Techpedia: A Glossary of Essential Tech Jargons

Data Governance

Data governance is the overall management of the availability, usability, integrity, and security of the data employed in an organization. It involves establishing policies, procedures, and standards for managing data, as well as monitoring compliance with those policies.

Data Modeling

Data modeling is the process of creating a representation of data, often in the form of a diagram or schema, that defines the structure and relationships of data elements within a system.

Data Mining

Data mining is the process of discovering patterns and knowledge from large amounts of data by using algorithms and statistical models.

Data Obfuscation

Data obfuscation is a data security strategy that replicates sensitive data and scrambles it, frequently using encryption, to hide it. Data masking and data obfuscation are frequently used interchangeably. Data is obfuscated to make it anonymous.

Digitization

Digitization refers to the process of converting analog information into a digital format. This can include scanning physical documents, photos, and videos into digital files, or converting analog audio or video recordings into a digital format. Digitization allows information to be stored, manipulated, and transmitted more efficiently and effectively.

Digitalization

Digitalization refers to the integration of digital technology into all areas of a business or organization. It involves using digital technology to transform processes, improve efficiency, and create new opportunities for growth and innovation. Digitalization goes beyond just digitizing information and involves a complete transformation of how a business operates.

API

API (Application Programming Interface) is an interface or code that enables two programmes to communicate with each other and use it resources. e.g. Twitter API, Google Maps API

Metaverse

The fusion of virtually augmented physical reality with virtualized digital reality results in the creation of a metaverse, a communal virtual 3D shared place. A persistent metaverse offers increased immersive experiences. The whole metaverse will be independent of hardware and not be controlled by a single company; it will have its own virtual economy made possible by digital money and non-fungible tokens (NFTs).

Quantum Computing

Quantum computing is a type of computing that uses the principles of quantum physics to perform calculations. In order to execute calculations, quantum computing makes advantage of the laws of quantum physics. Bits, which can be either a 0 or a 1, are used by conventional computers to store and process information. Quantum bits, or qubits, on the other hand, are used in quantum computers and can exist in several states at once. Qubits’ ability to do certain calculations substantially more quickly than conventional computers is a result of this qubit feature.

Quantum gates, which alter the qubits’ states, are used in a quantum computer to carry out operations. Quantum computers are able to solve issues that would be difficult or impractical for classical computers because these gates can be combined to carry out increasingly complicated operations.

Among the potential uses of quantum computing are optimization issues, drug discovery, and cryptography. However, quantum computers are still in the early stages of development and are currently limited in their capabilities.

Mechatronics

Mechatronics engineering is a multidisciplinary field of engineering that combines robotics, electronics, computer science, telecommunications, systems, control, and product engineering. It focuses on the integration of mechanical, electrical, and electronic engineering systems.
Interdisciplinarity

Ionic

Ionic Framework: Using well-known web libraries, frameworks, and languages, Ionic offers a collection of tools for developing native iOS and Android applications as well as mobile-ready Progressive Web Apps. Any online project may be converted into a native iOS or Android mobile application using the cross-platform native bridge known as Ionic Capacitor.

Tokenization

Tokenization is the process of breaking down a text into smaller units called tokens. In natural language processing (NLP), a token usually refers to a word, but it can also refer to a part of a word or punctuation mark.

Lemmatization

Lemmatization reduces words to their base form using a vocabulary and morphological analysis of the word. It takes into account the part of speech of the word and uses a dictionary to map the word to its base form, known as the lemma. In this case, the lemma of “desires” is “desire”.

Stemming

Stemming reduces words to a common stem or base form by removing suffixes. While stemming may sometimes result in the base form of a word, it is not as precise as lemmatization because it does not take into account the context or part of speech of the word.

NLP

Natural Language Processing (NLP) is a field of study that focuses on the interaction between human language and computers. It involves developing algorithms and models that can understand, interpret, and generate natural language, which is often ambiguous and context-dependent.

NLTK

NLTK (Natural Language Toolkit) is a popular Python library for NLP that provides tools and resources for tasks such as tokenization, stemming, lemmatization, part-of-speech tagging, and named entity recognition. It also includes a wide range of corpora and datasets for training and evaluation of NLP models. NLTK can be used for both research and production-level applications.

Neural Networks

Neural Networks are a type of machine learning model that are loosely inspired by the structure and function of the human brain. They consist of interconnected layers of nodes (neurons) that learn to perform complex tasks through a process of training on labeled data. Neural Networks have been successfully applied to many NLP tasks such as language modeling, machine translation, sentiment analysis, and text classification.

In summary of above 3 topics, NLP is a field of study that deals with the interaction between human language and computers. NLTK is a Python library that provides tools and resources for NLP tasks, and Neural Networks are a type of machine learning model that can be used to solve NLP problems.

  1. Docker:
    • Docker is a platform for developing, shipping, and running applications in containers. Containers are lightweight, portable, and self-sufficient units that can hold an application and its dependencies. Docker simplifies the process of building, distributing, and managing applications, making it easier to deploy software across different environments.
  2. Docker Compose:
    • Docker Compose is a tool for defining and running multi-container Docker applications. It allows you to define your application’s services, networks, and volumes in a single YAML file, making it easier to manage complex, multi-container applications. With Docker Compose, you can start, stop, and manage the entire application stack with a single command.
  3. Docker Hub:
    • Docker Hub is a cloud-based registry service provided by Docker for sharing container images. It’s a central repository where developers can publish and share Docker images. Users can easily pull and run container images from Docker Hub, which is especially helpful for finding and using pre-built images for various software and services.
  4. Docker Engine:
    • Docker Engine is the core component of Docker that runs and manages containers. It includes the Docker daemon, which is responsible for building, running, and managing containers. Docker Engine also provides a command-line interface for interacting with containers and images.
  5. Kubernetes:
    • Kubernetes is an open-source container orchestration platform for automating the deployment, scaling, and management of containerized applications. It allows you to define, deploy, and scale applications using container technology while providing features like load balancing, auto-healing, and resource management.
  6. K8 (Kubernetes):
    • “K8” is a shorthand term for “Kubernetes.” It’s derived by taking the first letter, the number of letters between, and the last letter in “Kubernetes.” It’s often used as an abbreviation to save time when typing or speaking about Kubernetes.
  7. WSL (Windows Subsystem for Linux):
    • WSL is a compatibility layer for running Linux binary executables natively on Windows 10 and Windows Server 2019. It allows developers and users to run a Linux distribution alongside their Windows installation, enabling them to use Linux tools and utilities without the need for a dedicated virtual machine or dual-boot setup.
  8. Ingress/Egress Traffic:
    • Ingress traffic refers to the data that flows into a network or a specific device within a network. It represents incoming data or requests from external sources, like users accessing a website or clients connecting to a server.
    • Egress traffic is the opposite of ingress traffic. It refers to the data leaving a network or a specific device within a network. Egress traffic usually includes responses to incoming requests, data sent from a server to a client, or any data leaving a network to reach external destinations.
  9. Load Balancing:
    • Load balancing refers to the practice of dynamically distributing network traffic across multiple servers or resources as needed. Load balancing helps ensure that no single server becomes overwhelmed with requests, which can improve performance, availability, and reliability of a service or application.
    • In an on-demand scenario, load balancing can automatically adjust the distribution of traffic based on real-time conditions, such as server health, traffic volume, and server capacity. This ensures that resources are efficiently utilized and that the system can handle sudden spikes in traffic.
  10. Prometheus:
    • Prometheus is an open-source monitoring and alerting toolkit. It is designed for collecting and storing time-series data, allowing you to monitor the performance and health of various systems, applications, and services. Prometheus is known for its simplicity, reliability, and a flexible query language for analyzing data.
    • Prometheus is often used in DevOps and system administration to track metrics and generate alerts when predefined conditions are met. It can be integrated with various data sources and is particularly popular in containerized and cloud-native environments.
  11. Grafana:
    • Grafana is an open-source platform for data visualization and monitoring. It can be used in conjunction with tools like Prometheus to create interactive and customizable dashboards that display various metrics and data in a user-friendly way.
    • Grafana supports a wide range of data sources, including databases, monitoring systems, and cloud services. Users can design and share dashboards that provide insights into the performance and status of applications, infrastructure, and other systems. It’s a popular choice for visualizing and exploring time-series data.

These technologies play a significant role in modern software development and deployment, especially in the context of containerization and cloud-native applications.