Non-Technical Introduction to AI Fundamentals

Photo of Krystian Bergmann

Krystian Bergmann

Updated Nov 22, 2024 • 14 min read
ai

Artificial Intelligence (AI) is the most powerful technological tide to date, and it’s rapidly sweeping through all industries.

From the lens of Nikolai Kondratiev's theory of economic cycles, AI is poised to propel us into a new wave of prosperity. Similarly, AI is surpassing the 'peak of inflated expectations' in Gartner’s hype cycle and inching closer to widespread adoption.

William Gibson once said, “the future is already here – it's just not evenly distributed.” With AI, the most powerful applications are currently reserved for those with astronomical budgets and access to near-infinite computing power. But, thanks to restless efforts of researchers, developers and entrepreneurs, that is changing very quickly.

ChatGPT broke records in consumer adoption. It achieved one million users in just five days, and 100 million users two months after its launch. Generative AI, which includes ChatGPT, is on the verge of "crossing the chasm" from early adopters to early majority. The key to this transition lies in proving the utility of AI concepts, finding the right pricing models, and ensuring accessibility.

Whether you're in finance, healthcare, retail, or any other sector really, AI impacts or will impact your work in some way. But for those without a tech background, the jargon can often be overwhelming. This artificial intelligence tutorial aims to bridge that gap and offer a comprehensive look at AI fundamentals.

Essential terminology

Artificial intelligence

Artificial Intelligence, often abbreviated as AI, refers to the capability of a machine to imitate human-like behavior. Think of it as teaching machines to "think" and make decisions, somewhat like a human would. AI is the umbrella term that houses various applications and technologies, including machine learning and deep learning.

Machine learning

How do you teach a computer to play checkers? This was actually one of the first working applications of machine learning algorithms, developed all the way back in 1959 by Arthur Samuel.

Machine learning is essentially a subset of AI that allows a computer system to learn from experience. In simpler terms, machine learning is about using data to answer questions. For instance, a machine learning algorithm might use sales data to predict next month's revenue.

Machine learning models work by:

  1. training a model on a set of data,
  2. using that model to make predictions or decisions.
This is unlike traditional programming, where every single rule must be explicitly programmed. Machine learning automates this by learning from data.

Example applications of machine learning:

  • Fraud detection in finance
  • Recommender systems for e-commerce
  • Predictive maintenance in manufacturing
  • Natural language processing in customer service
  • Image recognition in healthcare
  • Autonomous vehicles in transportation
  • Chatbots in customer relations
  • Speech recognition in voice-operated systems

Deep learning

Deep learning takes machine learning a step further. It employs more complex algorithms and uses an AI concept called deep neural networks to train the model. In a few words, how deep learning works is similar to how the human brain operates, with neurons, layers, and so on.

Essentially, deep learning can make sense of even more complex data sets, making it highly valuable for tasks that require a deeper level of data comprehension.

Example applications of deep learning:

  • Image and speech recognition
  • Natural language processing
  • Fraud detection
  • Recommender systems
  • Autonomous vehicles
  • Medical diagnosis
  • Financial modeling

Computer vision

Vision systems in AI are designed to interpret and understand visual data from the world, such as pictures and videos. They can identify objects, read signs, and more.

Applications of vision systems:

  • Object detection
  • Facial recognition
  • Image and video analysis
  • Autonomous vehicles
  • Medical imaging
  • Quality inspection
  • Surveillance

Natural language processing

Natural language processing (NLP) is a branch of AI that gives machines the ability to read, understand, and generate human language. This technology powers chatbots, virtual assistants, and other services that interact with people in a natural way.

Applications of NLP:

  • Chatbots and virtual assistants
  • Sentiment analysis
  • Language translation
  • Speech recognition
  • Text summarization
  • Customer service

Different Types of Algorithms

Regression algorithms

These are used for predicting numerical values, such as house prices. Examples include Linear regression, K-NN, and random forest.

Classification algorithms

These are used to classify data into different classes, like diagnosing a disease. Examples are Logistic regression, random forest, and K-NN.

Clustering algorithms

These categorize unlabelled data into similar groups. Examples are K-Means and DBSCAN.

Time-series forecasting algorithms

These predict future values based on past time-series data. Examples include ARIMA and LSTM.

Anomaly detection algorithms

These find outliers or unexpected items in datasets, often used in fraud detection. Examples are Isolation Forest and Local Outlier Factor.

Ranking algorithms

These order results based on certain criteria, such as recommendation engine outputs. Examples are Bipartite ranking methods like Bipartite Rankboost.

Recommendation algorithms

These recommend next items, like songs or videos. They generally use Content-based and collaborative filtering methods.

Data generation algorithms

These generate new data, like images or videos. Examples include LLM and Generative Adversarial Networks (GANs).

Optimization algorithms

These are used to find the best solution to a given problem based on an objective function. Examples are linear programming methods and genetic programming.

Designing AI-Based Products

Data labeling and preparation

Before diving into AI project steps, it's crucial to note that data is at the heart of every AI initiative. In fact, over 80% of the time in AI projects is often spent on data labeling and preparation. So, before you even think of algorithms, remember that quality data is your foundational block.

How we do it at Netguru

In broad strokes, here’s our process for designing AI-based products:

  1. AI Exploration Workshops: These workshops aim to explore the AI landscape, focusing on the needs and limitations of your particular project.
  2. AI Solution Design Sprint: A fast-paced brainstorming session where a potential AI solution is conceptualized.
  3. Data Engineering: Data is gathered, cleaned, and prepared for the upcoming phases.
  4. Proof-of-Concept Building: Here, a simple version of the solution is created to validate its feasibility and effectiveness.
  5. AI MVP Implementation: Finally, a Minimum Viable Product (MVP) is built, using AI to solve the problem at hand.

The 7 processes of machine learning

Machine learning projects generally follow a seven-step process:

  1. Data Gathering: The initial phase where you collect the relevant data.
  2. Data Pre-Processing: Data is cleaned and transformed. This can involve removing outliers or filling in missing values.
  3. Choose Model: You select the most suitable machine learning model for your problem.
  4. Train Model: The pre-processed data is used to train the selected model.
  5. Test Model: The model's performance is evaluated using a separate data set.
  6. Tune Model: Adjustments are made to the model to improve its performance.
  7. Prediction: The model is used for making predictions on new data.

What affects training data quality?

Data quality is influenced by:

  • Process: Including operations, business rules, and communication protocols.
  • People: The experience and training of the workforce handling the data.
  • Tools: Labeling tools and platforms used.

Solving problems: design thinking vs machine learning

Design thinking emphasizes human-centered solutions. Its process involves steps like empathizing with your audience, defining their needs, brainstorming ideas, prototyping solutions, and testing them.

Machine Learning takes a data-driven, machine-centered approach. It starts with dissecting the needs into base components, then synthesizing a solution, brainstorming requirements, tuning the model, and validating its performance.

Both methodologies offer unique lenses through which problems can be tackled. Design thinking focuses on human needs, while machine learning concentrates on data-driven solutions.

AI model development lifecycle

The AI model development process can broadly be divided into three stages:

  • Design and build: This involves data acquisition and data cleansing.
  • Deploy and operationalize: At this stage, data is acquired again, annotated, and the model is validated.
  • Refine and optimize: The model undergoes adjustments and optimizations for better performance and efficiency.

Understanding large language models (LLMs)

What is an LLM?

Large language models (LLM) are AI-driven powerhouses that can generate human-like text based on patterns they've learned from large datasets. The most popular is GPT, which stands for Generative Pre-trained Transformer. GPT-4 from OpenAI is the most powerful LLM as of this moment (20.10.2023), available for all to use through ChatGPT premium.

The swiss army knife of tech: LLM use-cases

LLMs can wear many hats. They can draft emails, answer questions about documents, act as a conversational agent, tutor in various subjects, translate languages, simulate video game characters, write Python code, and even give your software a natural language interface. In short, it can be the go-to tool for a multitude of tasks.

Attention: the secret LLM sauce

Why do people mention attention when discussing LLMs? The "attention mechanism" allows these models to focus on multiple parts of the input at once. Unlike their predecessors, such as LSTM or GRU, they do not rely on recurrent neural networks.

Thanks to features like multi-head attention, they can process different data types simultaneously. The architecture scales beautifully and is highly versatile, successfully tackling various natural language tasks.

The attention mechanism comes from the landmark paper “Attention is All You Need”, produced in 2017 by Google engineers who later went on to work in or found their own AI startups.

Proceed with caution: LLM limitations

LLMs can sometimes generate biased or outright false information. They may inadvertently produce sexist, racist, or homophobic responses. They also tend to falter with mathematical problems. Therefore, exercise caution when employing LLMs, especially in high-risk environments.

Cloud vs local LLMs

When it comes to hosting, you have two broad choices:

  • Proprietary cloud models like GPT-4 or GPT-3.5 Turbo offer scalability and ease of use but may compromise on control and data privacy.
  • Open-source local models like Llama from Meta give you greater control, privacy, and customizability but might require more work to get them to do what you want.

Each has its own set of pros and cons, so your decision should be tailored to your specific needs and constraints.

The horizon ahead: AI megatrends

The rise of co-pilots supporting human intelligence in everyday life

AI will be your new sidekick, no longer just recommending what movie to watch or product to buy, but aiding in complex daily tasks and decisions.

Code, but make it easier: genAI low-code

Low-code platforms powered by AI are making it simpler than ever to create robust applications, even if you're not a programming wizard.

The artist in the machine: generative creativity

Expect AI to eventually compose symphonies, create jaw-dropping art, and even write novels that tug at your heartstrings.

Democracy in machine learning

The tools for machine learning are becoming more accessible, letting even the tech-novices dip their toes in AI waters. Models are getting smaller and cheaper to run, and most AI research is open and available to all.

The new age of search and shopping

AI algorithms are revolutionizing how we search for information and shop online, making both experiences incredibly personalized.

The semiconductor gold rush

Companies aren't just consumers of AI; they're becoming manufacturers too, especially in the race for semiconductors. Microsoft, OpenAI, and Google among others are either working on, or already using their own AI hardware.

The global AI chess game: regulation

The European Union, the United States, and China are the biggest players grappling to set the rules of the AI game, influencing its development and adoption globally.

AI's influence on influencers

Your next Instagram influencer might just be a deepfake or a bot. In China, hundreds of thousands of deepfaked livestreamers are already raking in huge amounts of money in sales.

The classroom of tomorrow: education 2.0

AI is set to revolutionize how we learn, from personalized lesson plans to instant feedback. If you’re interested in learning more about how AI will change education, look no further than Professor Ethan Mollick, who has been studying this topic for a long time and has a rational, science-backed perspective.

Summary

Artificial Intelligence is reshaping how we solve problems, communicate, and even envision the future. More than just an innovative technology, it's a critical asset that's central to driving business success in an increasingly competitive landscape.

With a myriad of applications, from product design to natural language processing, the potential of AI is vast and still largely untapped. As we navigate the complexities and possibilities, it becomes ever more important to partner with experts who can guide us through this evolving landscape.

Netguru can help you appraise, design, and develop your AI ideas. Don't wait for the future, it’s time to accelerate!

Photo of Krystian Bergmann

More posts by this author

Krystian Bergmann

AI Consulting Lead at Netguru
Optimize with AI solutions  Automate processes and enhance efficiency   Get Started!

Read more on our Blog

Check out the knowledge base collected and distilled by experienced professionals.

We're Netguru

At Netguru we specialize in designing, building, shipping and scaling beautiful, usable products with blazing-fast efficiency.

Let's talk business