Sadman Kabir Soumik
open-menu closeme
Home
artificial-intelligence
software-engineering
cloud-computing
beyond-code
About
linkedin github rss
  • So You Want to Become a Machine Learning Engineer/Data Scientist?

    calendar Apr 19, 2025 · 7 min read · machine learning data science artificial-intelligence  ·
    Share on: twitter facebook linkedin copy
    So You Want to Become a Machine Learning Engineer/Data Scientist?

    I'm a fresh graduate and I want to become a data scientist or machine learning engineer. Can you please give me some guidance? I've been working as a software engineer for the last two years, but now I want to switch to data science. Can you help or share some guidance? I've been applying to many data science jobs but …


    Read More
  • Types of LLM Architectures

    calendar Mar 18, 2025 · 6 min read · machine learning data science artificial-intelligence nlp system-design  ·
    Share on: twitter facebook linkedin copy
    Types of LLM Architectures

    Let's first break it down: what exactly are large language models (LLMs), why do we call them 'large,' and how are they different from other types of language models? An LLM is a machine learning model trained on massive amounts of text using transformer-based architectures (or their variations). These models can …


    Read More
  • From RNN to Transformers (Without Math Jargon)

    calendar Jan 30, 2023 · 13 min read · machine learning data science NLP algorithm  ·
    Share on: twitter facebook linkedin copy
    From RNN to Transformers (Without Math Jargon)

    Transformer-based models are a types of neural network architecture that uses self-attention mechanisms to process input data. They were introduced in the paper "Attention Is All You Need" by Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia …


    Read More
  • Ace Your Data Science Interview - Top Questions With Answers

    calendar Nov 15, 2022 · 100 min read · deep learning machine learning data science  ·
    Share on: twitter facebook linkedin copy
    Ace Your Data Science Interview - Top Questions With Answers

    Can you explain the bias-variance trade-off and how it relates to model performance? Machine learning and statistics have a fundamental concept that requires balancing the model's bias and variance, known as the bias-variance trade-off. These two types of errors can affect a model's performance. Bias, the first type of …


    Read More
  • Understanding Top 10 Classical Machine Learning Algorithms

    calendar Oct 26, 2022 · 42 min read · machine learning data science algorithms  ·
    Share on: twitter facebook linkedin copy
    Understanding Top 10 Classical Machine Learning Algorithms

    Before jumping into Deep Learning, one must know the classical/traditional Machine Learning algorithms, because understanding traditional machine learning algorithms can provide a strong foundation in machine learning concepts. These algorithms often involve simple, intuitive concepts that can be helpful in …


    Read More
  • Machine Learning Model Compression Techniques - Reducing Size and Improving Performance

    calendar Oct 10, 2022 · 6 min read · mlops optimization machine learning data science  ·
    Share on: twitter facebook linkedin copy
    Machine Learning Model Compression Techniques - Reducing Size and Improving Performance

    There are 4 main approaches you can consider for model compression. They are: Quantization Pruning Knowledge Distillation Low-Rank Factorization Quantization Quantization is the most general and commonly used model compression method. Quantization reduces a model’s size by using fewer bits to represent its parameters. …


    Read More
  • Understanding the Role of Data Normalization and Standardization in Machine Learning

    calendar Mar 12, 2022 · 2 min read · data science statistics machine learning  ·
    Share on: twitter facebook linkedin copy
    Understanding the Role of Data Normalization and Standardization in Machine Learning

    Why do we scale features? For machine learning, every dataset does not require feature scaling, and it is only needed when features have different ranges. For example, consider a data set containing two features, age(x1) and income(x2), where age ranges from 0–100, while income ranges from 0–20,000 and higher. Income …


    Read More
  • One-Stage vs Two-Stage Instance Segmentation

    calendar Mar 4, 2022 · 5 min read · image segmentation machine learning computer vision data science  ·
    Share on: twitter facebook linkedin copy
    One-Stage vs Two-Stage Instance Segmentation

    In computer vision, image segmentation refers to the process of dividing an image into distinct regions or segments, each corresponding to a different object or background. There are two main approaches to image segmentation: one-stage and two-stage. One-stage image segmentation methods aim to directly predict a …


    Read More
  • Machine Learning Practices - Research vs Production

    calendar Jan 10, 2022 · 5 min read · machine learning deep learning data science  ·
    Share on: twitter facebook linkedin copy
    Machine Learning Practices - Research vs Production

    There are several key differences between using machine learning for research and using it for production. One of the main differences is the focus of the work. Machine learning for research typically focuses on exploring new ideas and techniques, and on advancing the state of the art in the field. In contrast, machine …


    Read More
  • Writing Machine Learning Model - PyTorch vs. TF-Keras

    calendar Dec 9, 2021 · 4 min read · machine learning deep learning data science  ·
    Share on: twitter facebook linkedin copy
    Writing Machine Learning Model - PyTorch vs. TF-Keras

    PyTorch and Keras are both open-source deep learning frameworks, but they have some significant differences. PyTorch is a low-level framework that allows you to define your own computation graphs, while Keras is a high-level framework that provides a pre-defined set of layers and routines for building deep learning …


    Read More
  • GPT-3 by OpenAI - The Largest and Most Advanced Language Model Ever Created

    calendar Nov 20, 2021 · 5 min read · machine learning NLP data science  ·
    Share on: twitter facebook linkedin copy
    GPT-3 by OpenAI - The Largest and Most Advanced Language Model Ever Created

    Author: Sadman Kabir Soumik GPT-3, or Generative Pretrained Transformer 3, is a state-of-the-art language model developed by OpenAI. It has been trained on a massive amount of text data, including books, articles, and websites, to generate coherent and relevant text based on a given context. GPT-3 is a …


    Read More
  • Vanishing Gradient Problem and How to Fix it

    calendar Oct 24, 2021 · 3 min read · deep learning machine learning data science  ·
    Share on: twitter facebook linkedin copy
    Vanishing Gradient Problem and How to Fix it

    What is Vanishing Gradient Problem Neural networks are trained using stochastic gradient descent. This involves first calculating the prediction error made by the model and using the error to estimate a gradient used to update each weight in the network so that less error is made next time. This error gradient is …


    Read More
  • Ensemble Techniques in Machine Learning - A Practical Guide to Bagging, Boosting, Stacking, Blending, and Bayesian Model Averaging

    calendar Jun 10, 2021 · 11 min read · machine learning data science  ·
    Share on: twitter facebook linkedin copy
    Ensemble Techniques in Machine Learning - A Practical Guide to Bagging, Boosting, Stacking, Blending, and Bayesian Model Averaging

    There are several types of ensemble techniques in machine learning, including: Bagging, Boosting, Stacking, Blending, Bootstrapped ensembles, Bayesian model averaging. Bagging Bagging (short for bootstrapped aggregating) is an ensemble technique that involves training multiple models on different subsets of the …


    Read More
  • Understanding the Differences between Decision Tree, Random Forest, and Gradient Boosting

    calendar Mar 27, 2021 · 4 min read · machine learning algorithm data science  ·
    Share on: twitter facebook linkedin copy
    Understanding the Differences between Decision Tree, Random Forest, and Gradient Boosting

    Decision Tree, Random Forest (RF), and Gradient Boosting (GB) are three popular algorithms used for supervised learning tasks such as classification and regression. In this blog, we will compare these three algorithms in terms of their features, performance, and usability. Decision Tree is a simple and intuitive …


    Read More
  • Different Text Cleaning Methods for NLP Tasks

    calendar May 4, 2020 · 5 min read · python programming NLP data science  ·
    Share on: twitter facebook linkedin copy
    Different Text Cleaning Methods for NLP Tasks

    Cleaning text for natural language processing (NLP) tasks is an important step that can help improve the performance of your model. In this blog post, we will discuss some common text cleaning techniques and how to apply them to your text data. The first step in cleaning text for NLP is to remove any noisy or …


    Read More
  • Different Types of Recommendation Systems

    calendar Oct 5, 2019 · 5 min read · machine learning data science  ·
    Share on: twitter facebook linkedin copy
    Different Types of Recommendation Systems

    There are several different types of recommender systems, each with its own unique characteristics and applications. Some of the most commonly used types of recommender systems include: Content-based recommender systems: These systems recommend items to users based on the characteristics of the items themselves. For …


    Read More
  • Effective Transfer Learning - A Guide to Feature Extraction and Fine-Tuning Techniques

    calendar May 21, 2018 · 4 min read · deep learning machine learning data science  ·
    Share on: twitter facebook linkedin copy
    Effective Transfer Learning - A Guide to Feature Extraction and Fine-Tuning Techniques

    Transfer learning is a technique in machine learning that allows a model trained on one task to be reused and fine-tuned for another similar task. The idea behind transfer learning is that a model that has already learned to recognize patterns in one set of data can be applied to a different but related problem, …


    Read More

SK Soumik photo

SK Soumik

Data Science | Software Engineering
Read More

Categories

MACHINE-LEARNING 25 SOFTWARE-ENGINEERING 18 ALGORITHM-DESIGN 10 SYSTEM-DESIGN 7 PROJECT 6 NON-TECHNICAL 2 OPERATING-SYSTEM 2 BUSINESS 1 CLOUD-COMPUTING 1

Series

ML 19 DSA 6 PROJECT 6 SOFTWARE-ENGINEERING 4 PROBLEM-SOLVING 3 AIRFLOW 1 DJANGO 1 GENERATIVE-AI 1 NLP 1

Tags

MACHINE-LEARNING 23 DATA-SCIENCE 17 PROGRAMMING 12 PYTHON 10 ALGORITHMS 8 DEEP-LEARNING 8 NLP 8 PROBLEM-SOLVING 6 PROJECT-TUTORIAL 6 SOFTWARE-ENGINEERING 6 ALGORITHM 5 LEETCODE 5 SYSTEM-DESIGN 5 ARTIFICIAL-INTELLIGENCE 3
All Tags
AIRFLOW1 ALGORITHM5 ALGORITHMS8 ARTIFICIAL-INTELLIGENCE3 AUTOMATION1 BOT-DEVELOPMENT1 BUSINESS1 CLOUD1 COMPUTER-VISION2 DATA-SCIENCE17 DEEP-LEARNING8 DJANGO1 ELASTICSEARCH1 GENERATIVE-AI1 GENERATIVE-AI-MODELS1 GOOGLE-CLOUD1 IMAGE-SEGMENTATION1 INDEX1 LEETCODE5 LIFE-HACK1 LINUX2 MACHINE-LEARNING23 MATH1 MLOPS2 NLP8 OPERATING-SYSTEM1 OPTIMIZATION1 PROBLEM-SOLVING6 PRODUCTIVITY1 PROGRAMMING12 PROJECT-TUTORIAL6 PROXY1 PYTHON10 SOFTWARE-ENGINEERING6 SPARK1 START-UP1 STATISTICS1 SYSTEM-DESIGN5 WEB-DEVELOPMENT1 WEB-SCRAPING1
[A~Z][0~9]
Copyright © 2022, Sadman Kabir Soumik

Copyright  COPYRIGHT © 2022, SADMAN KABIR SOUMIK. All Rights Reserved

to-top