Article Categories
- All Categories
-
Data Structure
-
Networking
-
RDBMS
-
Operating System
-
Java
-
MS Excel
-
iOS
-
HTML
-
CSS
-
Android
-
Python
-
C Programming
-
C++
-
C#
-
MongoDB
-
MySQL
-
Javascript
-
PHP
-
Economics & Finance
Articles on Trending Technologies
Technical articles with clear explanations and examples
Why do we need KDD?
The traditional techniques of turning data into knowledge depend on manual analysis and interpretation. For instance, in the healthcare industry, it is familiar for specialists to systematically analyze current trends and changes in healthcare data, every quarter.The specialists support a report detailing the analysis to the sponsoring healthcare organization; this report becomes the basis for future decision making and planning for healthcare management. There are several types of applications, including planetary geologists sifting through remotely sensed images of planets and asteroids, carefully situating and cataloging such geologic objects of interest as impact craters.This form of manual probing of a data ...
Read MoreWhat are Data Warehouse Users?
Data Warehousing is a technique that is generally used to collect and manage data from multiple sources to provide the business a meaningful business insight. A data warehouse is specifically created for the goals of support management decisions.In simple terms, a data warehouse defines a database that is maintained independently from an organization’s operational databases. Data warehouse systems enable the integration of several application systems. They provide data processing by supporting a solid platform of consolidated, historical data for analysis.There are various types of data warehouse users which are as follows −Statisticians − There are generally only a handful of ...
Read MoreDifference between a data warehouse database and an OLTP database?
Data Warehouse DatabaseData Warehousing is a technique that is generally used to collect and manage data from multiple sources to provide the business a meaningful business insight. A data warehouse is specifically created for the goals of support management decisions.In simple terms, a data warehouse defines a database that is maintained independently from an organization’s operational databases. Data warehouse systems enable the integration of several application systems. They provide data processing by supporting a solid platform of consolidated, historical data for analysis.A data warehouse provides an OLTP system by supporting a place for the OLTP database to offload data as ...
Read MoreWhat is the design of quality driven data warehouse?
A data warehouse defines a database that is maintained independently from an organization’s operational databases. Data warehouse systems enable the integration of several application systems. They support data processing by supporting a solid platform of consolidated, historical records for analysis.A data warehouse can be viewed as a set of materialized views represented over remote base relations. When a query is formal, it is computed locally, using the materialized views, without accessing the initial data sources.The data warehouse is an active entity that derives continuously over time. As time passes, new queries are required to be answered by them. Various queries ...
Read MoreNode.js – dns.resolve4() Method
The dns.resolve4() method uses the DNS protocol to resolve a IPv4 address for the hostname. The arguments passed to the callback function can contain an array of multiple addresses.Syntaxdns.resolve4(hostname, [options], callback)Parametershostname - This parameter takes input for the hostname to be resolved.options - It can have the following optionsttl - It defines the Time-To-Live (TTL) for each record. Callback receives an array of addresses like this{ address: '1.2.3.4', ttl:60 }callback - It will catch errors, if any.Example 1Create a file with the name "resolve4.js" and copy the following code snippet. After creating the file, use the command "node resolve4.js" to ...
Read MoreWhich function of scipy.cluster.vq module is used to normalize observations on each feature dimension?
Before implementing k-means algorithms, it is always beneficial to rescale each feature dimension of the observation set. The function scipy.cluster.vq.whiten(obs, check_finite = True)is used for this purpose. To give it unit variance, it divides each feature dimension of the observation by its standard deviation (SD).ParametersBelow are given the parameters of the function scipy.cluster.vq.whiten(obs, check_finite = True) −obs− ndarrayIt is an array, to be rescaled, where each row is an observation, and the columns are the features seen during each observation. The example is given below −obs = [[ 1., 1., 1.], [ 2., 2., 2.], ...
Read MoreHow can we call the documentation for NumPy and SciPy?
If you are unsure of how to use a particular function or variable in NumPy and SciPy, you can call for the documentation with the help of ‘?’. In Jupyter notebook and IPython shell we can call up the documentation as follows −ExampleIf you want to know NumPy sin () function, you can use the below code −import numpy as np np.sin?OutputWe will get the details about sin() function something like as follows −We can also view the source with the help of double question mark (??) as follows −import numpy as np np.sin??Similarly, if you want to see the ...
Read MoreTo work with SciPy, do I need to import the NumPy functions explicitly?
When SciPy is imported, you do not need to explicitly import the NumPy functions because by default all the NumPy functions are available through SciPy namespace. But as SciPy is built upon the NumPy arrays, we must need to know the basics of NumPy.As most parts of linear algebra deals with vectors and matrices only, let us understand the basic functionalities of NumPy vectors and matrices.Creating NumPy vectors by converting Python array-like objectsLet us understand this with the help of following example−Exampleimport numpy as np list_objects = [10, 20, 30, 40, 50, 60, 70, 80, 90] array_new = np.array(list_objects) print ...
Read MoreWhat is Data Staging?
In the data warehousing process, the data staging area is collected of the data staging server software and the data store archive (repository) of the results of extraction, transformation, and loading activity.The data staging software server temporarily saves and changes data extracted from OLTP data sources and the archival repository stores cleaned, transformed data and attributes for loading into data marts and data warehouses.The data staging process imports information either as streams or files, change it, produces integrated, cleaned data, and stages it for loading into data warehouses, data marts, or Operational Data Stores.A data staging tool is accessible, and ...
Read MoreWhat is the structure of the data warehouse?
Data Warehousing is a method that is generally used to collect and handle data from various other sources to provide the business a meaningful business insight. A data warehouse is specifically created for the goals of support management decisions. The Data Warehouse has two main parts which are as follows −Physical store − A Microsoft SQL Server database that it can query using SQL queries, and an OLAP database that it can need to run reports.Logical schema − A conceptual model that maps to the data in the physical store.Physical StoreThe physical store for the Data Warehouse contains one database ...
Read More