Skip to main content

Vprocs


DEFINITION

Vproc, acronym for "Virtual PROCessor," is a self-contained instance of the processor software on a Teradata platform (SMP platform or a node). Vprocs are the basic units of parallelism that make up the Teradata database server.


OVERVIEW


To put it simply, a virtual processor is a simulated processor in a processing software system, or a software version of a dedicated physical processor. Each vproc uses a portion of the physical processor resource, and runs independently of other vprocs.  The vprocs co-existing within a node share a single memory pool - a collection of free memory in the node. The portion of memory allocated from the pool to one vproc will be returned to the pool after usage for use by other vprocs.

Vprocs are the basic units of parallelism that make up the Teradata database server. A single SMP node is made up of multiple vprocs, and a single Teradata MPP system is made up of multiple SMP nodes.  Vprocs run as multi-threaded processes to enable Teradata's parallel execution of tasks without using specialized physical processors.  They are labeled as "Virtual Processors" because they perform the similar processing functions as physical processors (CPU).

The concept of virtual processors is often utilized in the multithreading technology for parallel processing. Some examples are available from IBM Informix, HP POSIX and VM Host (HP-UX 11i) and SUN Solaris OS, though the concept is used differently in each case.

The concept of vprocs was introduced to Teradata to eliminate the need for proprietary technology in the form of physical processors. With Teradata V1 - the version used on DBC 1012, there were three types of physical processors:

    * The InterFace Processor(IFP) was responsible for the communication between the DBC and the HOST. Its components included parser, dispatcher, session controller, client interface and YNET interface.
    * The COmmunication Processor (COP) was the type of IFP that contained a gateway process for communication with hosts (DOS-PC/UNIX) via a network.     * The Access Module Processor (AMP) was the physical processing unit for all the Teradata database functions. Each AMP contained its own microprocessor, disk drive, file system, database software (Database Manager), Teradata Operating System (TOS), and YNET interface. In some sense, an AMP was a node.

The YNET Interconnection Network linked all the IFPs, COPs, and AMPs together with circuit boards and cables.

In Teradata V2, the above-mentioned physical processors (IFP, COP, and AMP) were replaced by Virtual PROCessors (vprocs), which were invented as an abstraction layer between the work-unit multithreading and the physical parallel processing system. [6] Logically, each vproc is a seperate AMP or PE instance within an SMP node. Physically, a vproc is a directly addressable collection of processes that allow data correction and fault tolerance. Teradata allocates its data rows across all AMP vprocs via hash partitioning. PE vprocs manage sessions, break work units down to multiple steps, make query plans and distribute the steps to the relevant AMP vprocs. The AMPs process the steps in parallel. Differently from Teradata V1, there may be multiple AMPs in a node.

To sum up, as Carrie Ballinger and Ron Fryer observe, "the VPROC is the fundamental unit of apportionment, and delivers basic query parallelism to all work in the system."


 
TYPES

There are two types of vprocs on the Teradata platform: AMP(Access Module Processor) and PE(Parsing Engine).

    * AMPs are the type of vprocs that contain a BYNET interface and perform database, file and disk functions.
    * PEs are the type of vprocs that handle session control, SQL parsing/optimizing, step generating and dispatching.



RELATED UTILITIES

Vproc Manager (vprocmanager)

Vproc Manager(vprocmanager) is used for the following administrative functions:
    * To check the status of the specific vproc or vprocs;
    * To alter the state of a vproc or the states for a series of vprocs;
    * To initialize and start the specific vproc - mostly an AMP vproc;
    * To initialize the file subsystem on the vdisk associated with a certain AMP vproc;
    * To force a manual database restart.

Comments

Popular posts from this blog

Informatica Powercenter Partitioning

Informatica PowerCenter Partitioning Option increases the performance of PowerCenter through parallel data processing. This option provides a thread-based architecture and automatic data partitioning that optimizes parallel processing on multiprocessor and grid-based hardware environments. Introduction: With the Partitioning Option, you can execute optimal parallel sessions by dividing data processing into subsets that are run in parallel and spread among available CPUs in a multiprocessor system. When different processors share the computational load,large data volumes can be processed faster. When sourcing and targeting relational databases, the Partitioning Option enables PowerCenter to automatically align its partitions with database table partitions to improve performance. Unlike approaches that require manual data partitioning, data integrity is automatically guaranteed because the parallel engine of PowerCenter dynamically realigns data partitions for set-oriented trans...

Data virtualization

Data virtualization is a process of offering a data access interface that hides the technical aspects of stored data, such as location, storage structure, API, access language, and storage technology. Analogous to concept of views in databases Data virtualization tools come with capabilities of  data integration, data federation, and data modeling Requires more memory caching Can integrate several data marts or data warehouses through a  single data virtualization layer This concept and software is a subset of data integration and is commonly used within business intelligence, service-oriented architecture data services, cloud computing, enterprise search, and master data management. Composite, Denodo, and Informatica are the largest players in the area of data virtualization References for definition: http://www.b-eye-network.com/view/14815

Find Changed Data by computing Checksum using MD5 function in Informatica

Introduction: Capturing and preserving the state of data across time is one of the core functions of a data warehouse, but CDC can be utilized in any database or data integration tool. There are many methodologies such as Timestamp, Versioning, Status indicators, Triggers and Transaction logs exists but MD5 function outlines on working with Checksum. Overview: MD5 stands for Message Digest 5 algorithm.It calculates the checksum of the input value using a cryptographic Message-Digest algorithm 5 and returns a128-bit 32 character string of hexadecimal digits (0 - F). Advantage of using MD5 function is that, it will reduce overall ETL run time and also reduces cache memory usage by caching only required fields which are utmost necessary. Implementation Steps : Identify the ports from the source which are subjected to change. Concatenate all the ports and pass them as parameter to MD5 function in   expression transformation Map the MD5 function output to a checksum outp...