Find Interview Questions for Top Companies
Ques:- What is NLS in datastage? how we use NLS in Datastage ? what advantages in that ? at the time of installation i am not choosen that NLS option , now i want to use that options what can i do ? to reinstall that datastage or first uninstall and install once
Asked In :- aptean, axi,
Right Answer:
NLS stands for National Language Support in DataStage, which allows for the handling of multiple languages and character sets. You can use NLS in DataStage to ensure that data is processed correctly in different languages and to support internationalization.

The advantages of using NLS include improved data handling for multilingual data, better user experience for non-English speakers, and compliance with global data standards.

If you did not choose the NLS option during installation, you typically need to uninstall DataStage and then reinstall it with the NLS option selected, as NLS cannot be added to an existing installation.
Ques:- How we can secure our data in local area Network
Right Answer:
To secure data in a local area network (LAN), you can implement the following measures:

1. Use strong passwords for all devices and accounts.
2. Enable network encryption (e.g., WPA3 for Wi-Fi).
3. Implement a firewall to monitor and control incoming and outgoing traffic.
4. Use Virtual Private Networks (VPNs) for secure remote access.
5. Regularly update software and firmware to patch vulnerabilities.
6. Limit access to sensitive data through user permissions and roles.
7. Use antivirus and anti-malware software to protect against threats.
8. Regularly back up data to prevent loss from attacks.
9. Monitor network traffic for unusual activity.
10. Educate users about security best practices and phishing threats.
Ques:- How we use NLS function in Datastage? what are advantages of NLS function? where we can use that one? explain briefly?
Right Answer:
In DataStage, the NLS (National Language Support) function is used to handle multilingual data and ensure proper character encoding. It allows you to convert data between different character sets and formats, making it easier to work with international data.

Advantages of the NLS function include:
1. **Character Set Conversion**: It enables conversion between various character sets, ensuring data integrity across different languages.
2. **Locale-Specific Formatting**: It allows formatting of data according to specific regional settings, such as date and number formats.
3. **Improved Data Quality**: By correctly handling different languages and characters, it helps maintain data quality and accuracy.

You can use the NLS function in scenarios where you need to process or transform data that includes multiple languages or requires specific formatting based on locale.
Ques:- 1.What about System variables? 2.How can we create Containers? 3.How can we improve the performance of DataStage? 4.what are the Job parameters? 5.what is the difference between routine and transform and function? 6.What are all the third party tools used
Right Answer:
1. System variables are predefined variables in DataStage that provide information about the job environment, such as job name, job ID, and user name.

2. Containers can be created in DataStage by using the "Container" stage, which allows you to group multiple stages into a single reusable component.

3. Performance of DataStage can be improved by optimizing job design, using parallel processing, minimizing data movement, and tuning database queries.

4. Job parameters are variables that can be set at runtime to pass values into DataStage jobs, allowing for dynamic job execution.

5. A routine is a piece of code that can be reused across jobs, a transform is a specific operation applied to data, and a function is a predefined operation that returns a value.

6. Common third-party tools used with DataStage include IBM InfoSphere Data Quality, IBM InfoSphere Information Analyzer, and ETL tools like Talend or Informatica.
Ques:- How can I extract data from DB2 (on IBM iSeries) to the data warehouse via Datastage as the ETL tool. I mean do I first need to use ODBC to create connectivity and use an adapter for the extraction and transformation of data? Thanks so much if anybody cou
Asked In :- OCR Services, smartanalyst,
Right Answer:
Yes, you need to use ODBC to create connectivity to the DB2 database on IBM iSeries. After establishing the connection, you can use DataStage to extract and transform the data using the appropriate adapters for DB2.
Ques:- My requirement is like this :Here is the codification suggested: SALE_HEADER_XXXXX_YYYYMMDD.PSVSALE_LINE_XXXXX_YYYYMMDD.PSVXXXXX = LVM sequence to ensure unicity and continuity of file exchangesCaution, there will an increment to implement.YYYYMMDD = LVM
Asked In :- knapp ag, bloomreach, caas,
Right Answer:
To implement the required codification for your files, you should name them as follows:

- For sale header files: `SALE_HEADER_XXXXX_YYYYMMDD.PSV`
- For sale line files: `SALE_LINE_XXXXX_YYYYMMDD.PSV`

Where `XXXXX` is the LVM sequence for uniqueness and continuity, and `YYYYMMDD` represents the date in the format of year, month, and day.
Ques:- How i create datastage Engine stop start script.Actually my idea is as below.!#bin/bashdsadm – usersu – rootpassword (encript)DSHOMEBIN=/Ascential/DataStage/home/dsadm/Ascential/DataStage/DSEngine/binif check ps -ef | grep DataStage (client connection is
Asked In :- Neo Code,
Right Answer:
```bash
#!/bin/bash

# Set DataStage environment variables
DSHOME=/Ascential/DataStage/home/dsadm/Ascential/DataStage
export DSHOME
PATH=$DSHOME/DSEngine/bin:$PATH
export PATH

# Function to start DataStage Engine
start_engine() {
if ! ps -ef | grep -v grep | grep -q DataStage; then
echo "Starting DataStage Engine..."
su - dsadm -c "$DSHOME/DSEngine/bin/dsadm start"
else
echo "DataStage Engine is already running."
fi
}

# Function to stop DataStage Engine
stop_engine() {
if ps -ef | grep -v grep | grep -q DataStage; then
echo "Stopping DataStage Engine..."
su - dsadm -c "$DSHOME/DSEngine/bin/dsadm stop"
else
echo "DataStage Engine is not running."
Ques:- Can any one tell me how to extract data from more than 1 hetrogenious Sources.mean, example 1 sequenal file, Sybase , Oracle in a singale Job.
Right Answer:
To extract data from multiple heterogeneous sources like a sequential file, Sybase, and Oracle in a single job, you can use an ETL (Extract, Transform, Load) tool such as Apache Nifi, Talend, or Informatica. These tools allow you to connect to various data sources, perform necessary transformations, and load the data into a target system in one workflow. You can configure the job to read from each source, process the data as needed, and then combine or store the results in a unified format.
Ques:- What other ETLs you have worked with?
Asked In :- jungleworks,
Right Answer:
I have worked with ETL tools such as Apache NiFi, Talend, Informatica, Microsoft SQL Server Integration Services (SSIS), and AWS Glue.
Ques:- * What are constraints and derivation?* Explain the process of taking backup in DataStage?*What are the different types of lookups available in DataStage?
Right Answer:
**Constraints and Derivation:**
- **Constraints** are rules that limit the type of data that can be entered into a database, ensuring data integrity (e.g., primary key, foreign key, unique, check constraints).
- **Derivation** refers to the process of calculating or transforming data values based on existing data, often using expressions or formulas in DataStage.

**Backup Process in DataStage:**
1. Stop the DataStage services.
2. Backup the DataStage project directories and configuration files.
3. Backup the DataStage repository database (e.g., using database-specific backup tools).
4. Restart the DataStage services after the backup is complete.

**Types of Lookups in DataStage:**
1. **Static Lookup**: Uses a fixed set of data for lookups.
2. **Dynamic Lookup**: Allows for real-time data retrieval and can update the lookup table.
3. **Sparse Lookup**: Efficiently handles large datasets by only loading relevant data into memory.
Ques:- Where does unix script of datastage executes weather in clinet machine or in server.suppose if it eexcutes on server then it will execute ?
Asked In :-
Right Answer:
The Unix script of DataStage executes on the server.
Ques:- How can u implement slowly changed dimensions in datastage? explain?2) can u join flat file and database in datastage?how?
Asked In :- dataeaze systems, tricentis,
Right Answer:
1) To implement slowly changing dimensions (SCD) in DataStage, you can use the SCD stage, which allows you to manage changes in dimension data over time. You can configure it to handle different types of SCDs (Type 1, Type 2, Type 3) by specifying how to update existing records, insert new records, or maintain historical data.

2) Yes, you can join a flat file and a database in DataStage by using the Join stage. You can read data from the flat file using a Sequential File stage and from the database using a Database stage, then connect both to the Join stage to perform the join operation based on a common key.
Ques:- How can I connect my DB2 database on AS400 to DataStage? Do I need to use ODBC 1st to open the database connectivity and then use an adapter for just connecting between the two? Thanks alot of any replies.
Asked In :- Finacle Technologies, enoc,
Right Answer:
To connect your DB2 database on AS400 to DataStage, you can use the IBM DB2 Connector for DataStage, which allows direct connectivity. You do not necessarily need to use ODBC first; the DB2 Connector can handle the connection directly. Ensure you have the appropriate drivers and configurations set up for the connection.
Ques:- What happends out put of hash file is connected to transformer ..What error it throughs
Right Answer:
If the output of a hash file is connected to a transformer, it will throw an error indicating that the hash file is not a valid source for the transformer, as transformers require a relational source or a sequential file.
Ques:- What is DS Director used for – did u use it?
Asked In :-
Right Answer:
DS Director is used for managing and monitoring DataStage jobs and workflows in IBM InfoSphere DataStage. It provides a graphical interface for job design, execution, and scheduling.
Ques:- How much would be the size of the database in DataStage ?What is the difference between Inprocess and Interprocess ?
Asked In :-
Right Answer:
The size of the database in DataStage depends on the amount of data being processed and stored, as well as the design of the database schema.

The difference between Inprocess and Interprocess is that Inprocess refers to operations that occur within the same process or memory space, while Interprocess involves communication and data exchange between different processes or memory spaces.
Ques:- How to sort the data in Datagrid ? when we use DataBound event of Datagrid, I want to display te data using controls….For an example, If i click a button, which is outside of the datagrid, entire data should be shown. 1st column in TextBoxs. second colu
Asked In :- andela, infocepts,
Right Answer:
To sort data in a DataGrid and display it in TextBoxes when a button is clicked, you can follow these steps:

1. Handle the button click event.
2. Retrieve the data from the DataGrid's data source.
3. Sort the data programmatically (e.g., using LINQ or sorting methods).
4. Populate the TextBoxes with the sorted data.

Here's a simple example in C#:

```csharp
private void Button_Click(object sender, EventArgs e)
{
// Assuming 'dataSource' is your data source for the DataGrid
var sortedData = dataSource.OrderBy(item => item.FirstColumn).ToList();

// Display sorted data in TextBoxes
textBox1.Text = sortedData[0].FirstColumn.ToString();
textBox2.Text = sortedData[0].SecondColumn.ToString();
// Repeat for other TextBoxes as needed
}
```

Make sure to replace `FirstColumn` and `SecondColumn`
Ques:- Consider a sales invoice and the programmer has to save multiple items in one sales invoice. On entering the rate and qty, the amount and total amount should be automatically calculated. The invoice number will be the same for all items in that invoice. I
Right Answer:
To design a database for a sales invoice with multiple items, create two tables:

1. **Invoices Table**:
- `InvoiceID` (Primary Key)
- `InvoiceDate`
- `CustomerID`
- `TotalAmount` (calculated as the sum of all item amounts)

2. **InvoiceItems Table**:
- `ItemID` (Primary Key)
- `InvoiceID` (Foreign Key referencing Invoices)
- `ItemDescription`
- `Rate`
- `Quantity`
- `Amount` (calculated as `Rate * Quantity`)

When a user enters the rate and quantity for each item, the amount is calculated and stored in the InvoiceItems table. The total amount for the invoice is calculated by summing the amounts of all related items in the InvoiceItems table.
Ques:- I wanted to load only particular no of records suppose i have 100 records out of which i wanted to load first 10 records to target.
Right Answer:
You can use the SQL `LIMIT` clause to load a specific number of records. For example, to load the first 10 records, you can use the following query:

```sql
SELECT * FROM your_table_name
LIMIT 10;
```
Ques:- What is data validation strategies for data mart
Right Answer:
Data validation strategies for a data mart include:

1. **Source Data Validation**: Ensure data from source systems is accurate and complete before loading.
2. **Transformation Validation**: Verify that data transformations during ETL processes are correct and maintain data integrity.
3. **Data Consistency Checks**: Regularly check for consistency across different data sources and within the data mart itself.
4. **Business Rule Validation**: Implement checks to ensure data adheres to defined business rules and logic.
5. **Data Profiling**: Analyze data for quality issues such as duplicates, missing values, and outliers.
6. **User Acceptance Testing (UAT)**: Involve end-users to validate that the data meets their requirements and expectations.
7. **Automated Testing**: Use automated tools to regularly test data quality and integrity throughout the data mart lifecycle.


The Database Architecture category on takluu.com is curated for professionals and aspirants seeking to master the principles and design strategies behind effective database systems. Good database architecture ensures data integrity, performance optimization, scalability, and security, which are vital for any data-driven application.

This section covers foundational concepts such as database models (relational, NoSQL, object-oriented), schema design, normalization, indexing, partitioning, and transaction management. You will also learn about physical and logical database design, data warehousing architecture, and distributed databases.

Interview questions often test your ability to design efficient database schemas, optimize queries, and ensure consistency and availability. Topics like ACID properties, CAP theorem, replication, sharding, and backup strategies are also commonly covered.

Candidates preparing for roles like Database Architect, Data Engineer, or Senior DBA will benefit from practical insights into designing scalable and maintainable database systems that meet business requirements.

At Takluu, we provide comprehensive study materials, real-world scenarios, and interview questions that help you build a strong foundation in database architecture principles. Understanding these concepts will enable you to design robust databases that can handle growing data and complex queries efficiently.

Whether you’re interviewing for a new role or looking to deepen your knowledge, this category equips you with the skills needed to excel in database design and architecture.

AmbitionBox Logo

What makes Takluu valuable for interview preparation?

1 Lakh+
Companies
6 Lakh+
Interview Questions
50K+
Job Profiles
20K+
Users