Machine Learning – AIS Home | Assured Information Security https://www.ainfosec.com Thu, 05 Dec 2024 20:06:29 +0000 en-US hourly 1 https://wordpress.org/?v=6.6.2 https://www.ainfosec.com/wp-content/uploads/2022/10/cropped-ais-icon-1-32x32.png Machine Learning – AIS Home | Assured Information Security https://www.ainfosec.com 32 32 5 Seminal Papers to Kickstart Your Journey Into Large Language Models https://www.ainfosec.com/5-seminal-papers-to-kickstart-your-journey-into-large-language-models?utm_source=rss&utm_medium=rss&utm_campaign=5-seminal-papers-to-kickstart-your-journey-into-large-language-models Tue, 22 Oct 2024 19:38:53 +0000 https://www.ainfosec.com/?p=20449 ...]]>
blogpic-bowman

Author: Dylan, Research Scientist II
Five minute read

Large language models (LLMs) have introduced an exciting new paradigm to the machine learning world in the past few years.

Research groups like AIS’s Advanced Research Concepts (ARC) team have been quick in exploring the range of possibilities with this new technology. However, aspiring AI developers and scientists often have no idea where to start learning about the science of large language models. In this blog post, we’ll review five seminal papers from the field and provide a brief reading guide so that you know which details to look for when you read them yourself.

1

Attention Is All You Need

“Attention Is All You Need” (Vaswani et al., 2017) presented the transformer architecture for neural networks which underlies all modern LLM designs. Transformers differed from previous neural networks in their use of self-attention, a mechanism that computes the ways elements in the input sequence affect each other’s semantic value. While the architecture was initially presented for use in language translation, it was adapted and scaled up by OpenAI into their first GPT model as well as Google’s T5 and BERT.

What Will This Paper Teach You?

  • The self-attention mechanism.
  • How the transformer architecture is organized into layers that update a residual stream.
  • Autoregressive language generation and how it fits into the sequence-to-sequence paradigm.

2

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

Another paper to come out of the early transformer literature, “BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding” (Devlin et al., 2019), set the paradigm for transformer embedding models, which seek to embed sequences of text into vectors. The BERT model uses the encoder part of the transformer architecture and is trained on unsupervised masked language modeling on a large set of unstructured text data. AIS used MiniLM, a successor to BERT, as part of an explanatory reinforcement learning technique in our paper “CODEX: A Cluster-Based Method for Explainable Reinforcement Learning”.

What Will This Paper Teach You?

  • How the encoder part of the transformer architecture can be isolated for embedding models.
  • The paradigm of pre-training and fine-tuning with language models.

3

Language Models are Few-Shot Learners

While all of OpenAI’s GPT papers have been influential, the paper covering GPT-3, “Language Models are Few-Shot Learners” (Brown et al., 2020) marked the entry into the modern era of LLMs. The paper reviews the training process and architecture of GPT-3, but mostly focuses on the efficacy of few-shot learning, a technique where a pre-trained LLM is fed examples into its context window instead of being fine-tuned for a task.

What Will This Paper Teach You?

  • How the decoder part of the transformer architecture can be isolated for pure autoregressive generation.
  • The advantages of few-shot learning over fine-tuning.
  • The test sets used to evaluate capabilities of large language models, like HellaSwag, Winograd schemas and TriviaQA.

4

ReAct: Synergizing Reasoning and Acting in Language Models

One area of focus within LLM research is using prompt scaffolding and reasoning schemes to elicit more powerful capabilities from LLMs. A core paper in this subfield is “ReAct: Synergizing Reasoning and Acting in Language Models” (Yao et al., 2023), which builds on top of prior chain-of-thought prompting to get the model to think clearly and also use resources like a Wikipedia API to inform its actions. The ReAct framework is deployed in direct QA tasks as well as text-based adventure environments and shows improvements over base model performance in both.

What Will This Paper Teach You?

  • Prompt scaffolding frameworks and the degree to which they augment LLM abilities.
  • The evaluation environments for LLM agents like ALFWorld and WebShop.

5

The Llama 3 Herd of Models

To get a better idea of how modern LLMs are trained, we recommend “The Llama 3 Herd of Models” (Dubey et al., 2024). The Llama models are among the best open-source models today alongside the Qwen series and Mistral series, and the AIS team uses Llama models for several of our internal projects. This paper goes into depth on the more technical engineering details that went into the training process, especially post-training where they apply several techniques like SFT and DPO to coerce the base model into being more coherent and user-interactive.

What Will This Paper Teach You?

  • The post-training process for modern LLMs, including alignment, multilinguality and coding specialization.
  • How scaling laws are used during training to optimize energy expenditure (specifically, the data size/model size tradeoff in the compute budget).
  • The hardware and software used to support the platforms for training LLMs.
While these papers should provide a solid foundation to any novice LLM enthusiast, reading will only get you so far.

It’s important to engage with these papers and others in a more hands-on way: re-implementing architectures, reproducing results and performing experiments on top of the prior work are all great ways to learn more about LLMs, and in machine learning more generally. In particular, papers that use frontier model APIs like the ReAct paper above don’t require massive GPU servers and can run on a consumer-grade laptop. We hope that the resources we’ve provided will give a nice jumpstart!

]]>
NEW AIS Research Paper: CODEX https://www.ainfosec.com/new-ais-research-paper-codex?utm_source=rss&utm_medium=rss&utm_campaign=new-ais-research-paper-codex Thu, 21 Dec 2023 19:13:11 +0000 https://www.ainfosec.com/?p=18881 ...]]>

CODEX: A Cluster-Based Method for Explainable Reinforcement Learning

AIS research scientists Timothy Mathes, PhD, and Andrés Colón have published a research paper titled CODEX: A Cluster-Based Method for Explainable Reinforcement Learning. This research was done in collaboration with Georgia Tech Research Institute and the Air Force Research Laboratory.

“Our research proposes a method for AI explainability that fuses techniques from Computer Vision and Natural Language Processing,” said Mathes. “We believe that investigating interdisciplinary approaches to understanding AI decision-making will be crucial for harnessing the power of intelligent systems.”

The team presented the paper during a virtual workshop at the International Joint Conference on Artificial Intelligence (IJCAI) Explainable AI Workshop on August 31, 2023.

Abstract:

Despite the impressive feats demonstrated by Reinforcement Learning (RL), these algorithms have seen little adoption in high-risk, real-world applications due to current difficulties in explaining RL agent actions and building user trust. We present Counterfactual Demonstrations for Explanation (CODEX), a method that incorporates semantic clustering, which can effectively summarize RL agent behavior in the state-action space. Experimentation on the MiniGrid and StarCraft II gaming environments reveals the semantic clusters retain temporal as well as entity information, which is reflected in the constructed summary of agent behavior. Furthermore, clustering the discrete+continuous game-state latent representations identifies the most crucial episodic events, demonstrating a relationship between the latent and semantic spaces. This work contributes to the growing body of work that strives to unlock the power of RL for widespread use by leveraging and extending techniques from Natural Language Processing.

Learn more about research at AIS: www.ainfosec.com/R&D

]]>
CODEX: A Cluster-Based Method for Explainable Reinforcement Learning https://www.ainfosec.com/rd/codex-a-cluster-based-method-for-explainable-reinforcement-learning/?utm_source=rss&utm_medium=rss&utm_campaign=codex-a-cluster-based-method-for-explainable-reinforcement-learning Thu, 21 Dec 2023 16:05:49 +0000 https://www.ainfosec.com/?page_id=18856 ...]]>

R&D  >  CODEX

RESEARCH & DEVELOPMENT

CODEX:

A Cluster-Based Method for Explainable Reinforcement Learning

OVERVIEW

Despite the impressive feats demonstrated by Reinforcement Learning (RL), these algorithms have seen little adoption in high-risk, real-world applications due to current difficulties in explaining RL agent actions and building user trust.

We present Counterfactual Demonstrations for Explanation (CODEX), a method that incorporates semantic clustering, which can effectively summarize RL agent behavior in the state-action space.

Experimentation on the MiniGrid and StarCraft II gaming environments reveals the semantic clusters retain temporal as well as entity information, which is reflected in the constructed summary of agent behavior. Furthermore, clustering the discrete+continuous game-state latent representations identifies the most crucial episodic events, demonstrating a relationship between the latent and semantic spaces. This work contributes to the growing body of work that strives to unlock the power of RL for widespread use by leveraging and extending techniques from Natural Language Processing.

Reach Out Today

Push your mission forward with the right partner.
]]>
Advancements in AI/ML Explainability and Safety Research https://www.ainfosec.com/advancements-in-ai-ml-explainability-and-safety-research?utm_source=rss&utm_medium=rss&utm_campaign=advancements-in-ai-ml-explainability-and-safety-research-at-ais Mon, 11 Dec 2023 17:59:52 +0000 https://www.ainfosec.com/?p=18749 ...]]>

In 2023, researchers at AIS proposed an internal research and development (IRaD) project seeking to develop novel technologies to deeply inspect Artificial Intelligence (AI) and Machine Learning (ML) based systems and algorithms. The team addresses this problem with a multi-disciplinary approach, applying knowledge from different internal groups at AIS including experts in AI/ML research and experts in the research and development of innovative software reverse engineering technologies. AIS’s AI/ML expertise spans the fields of Explainable AI (XAI), reinforcement learning, adversarial learning, safety and natural language processing.

“Before we can confidently rollout and leverage the power of AI/ML technologies, we need to better understand the algorithm’s process for achieving a certain conclusion,” said Georgia-Raye Moore, Advanced Research Program Manager. “This need has spawned into its own field within the industry and is referred to as ‘explainability’ or Explainable AI (XAI).”

AIS’s history of research in software analysis and inspection, which dates back to our founding, lends itself to an application in AI/ML based technologies.

“Decision makers need methods to deeply understand the risks and benefits of the AI/ML systems proposed for deployment within critical environments,” said Andrés Colón, Principal Investigator for the project. “The question our team seeks to address is: How can we better understand the underlying behaviors of often opaque AI/ML systems and build confidence prior to deployment?”

As part of this effort, AIS has been working to contribute to the space of XAI through the creation of a platform for unprecedented insight into AI/ML models.

“We are currently probing neural networks at the lowest levels during execution and surfacing this information to powerful analysis and visualization routines, clearing the path from opaque black-box neural network to unprecedented insight,” said Logan Boyd, Research Scientist and Technical Lead for the project. “The system is designed around extensibility so that researchers can build upon it and continue to push the state of the art in novel and interesting ways.”

The team’s preliminary assessment of existing AI/ML models has shown promise in the tool’s ability to identify, inspect and visualize a model’s inner workings, allowing us to gain valuable insights. The research suggests promising contributions to AI/ML analysis and assessment in areas such as quality assurance pipelines, extracting runtime information and visualization. AIS is currently leveraging this tool for innovative research in the fields of adversarial input detection, large language models and XAI.

The team is especially excited about this work as it lends to the testing and decomposition of complex AI/ML applications that are currently difficult to understand. Most importantly, this research supports making revolutionary AI/ML systems more trustworthy so they can confidently be deployed.

“It’s important for those adopting new AI/ML technologies to understand that models can have flaws, inherent limitations or blind spots,” said Gary Hamilton, Strategic Account Manager. “Our research aims to make these technologies safer by better understanding the system and enabling rapid adoption to ensure continued superiority.”

The team is eager to continue this research and the development of their tool in 2024. Learn more about advanced research at AIS: https://www.ainfosec.com/capabilities/advanced-research/

]]>
FISSURE https://www.ainfosec.com/rd/fissure-infosheet/?utm_source=rss&utm_medium=rss&utm_campaign=fissure Mon, 28 Nov 2022 20:35:31 +0000 https://www.ainfosec.com/?page_id=16685 ...]]>

R&D  >  FISSURE

RESEARCH & DEVELOPMENT

FISSURE

Frequency Independent SDR-Based Signal Understanding and Reverse Engineering 

OVERVIEW

FISSURE is an open-source RF and reverse engineering framework designed for all skill levels with hooks for signal detection and classification, protocol discovery, attack execution, IQ manipulation, vulnerability analysis, automation and AI/ML. FISSURE is a workflow enabler that keeps software in one location and allows teams to effortlessly get up to speed while sharing the same proven baseline configuration for specific Linux distributions.

KEY FEATURES
  • Detects the presence of RF energy
  • Understands the characteristics of a signal
  • Collects and analyzes samples
  • Develops transmit and/or injection techniques
  • Crafts custom payloads or messages

Reach Out Today

Push your mission forward with the right partner.
]]>
UAS https://www.ainfosec.com/rd/uas/?utm_source=rss&utm_medium=rss&utm_campaign=uas Mon, 28 Nov 2022 20:18:03 +0000 https://www.ainfosec.com/?page_id=16672 ...]]>

R&D  >  Cyber Support for UAS

RESEARCH & DEVELOPMENT

Cyber Support for UAS

Full Spectrum Cyber Support for Unmanned Aerial Systems 

OVERVIEW

In a world of evolving technology, we can help ensure that your Unmanned Aerial Systems (UAS) are secure. We help protect against attacks that alter intended operations and create significant safety issues. We also develop solutions to detect, monitor and protect from unwanted platforms.

KEY FEATURES
  • Consulting and design
  • Testing and evaluation
  • Secure solution development
  • cUAS capability development

Reach Out Today

Push your mission forward with the right partner.
]]>
BSL https://www.ainfosec.com/rd/bsl/?utm_source=rss&utm_medium=rss&utm_campaign=bsl Mon, 28 Nov 2022 20:03:12 +0000 https://www.ainfosec.com/?page_id=16657 ...]]>

R&D  >  BSL

RESEARCH & DEVELOPMENT

BSL

Bareflank Support Library  

OVERVIEW

The Bareflank Support Library (BSL) is a C++20, AUTOSAR and C++ Core Guideline compliant header-only library intended to support the development of critical systems applications using the Clang/LLVM compiler.

KEY FEATURES
  • Header only
  • AUTOSAR/Core guidelines compliance
  • C++ 20 support
  • Dynamic memory and expectations not required
  • 100% unit testing
  • Static analysis included

Reach Out Today

Push your mission forward with the right partner.
]]>
ARM https://www.ainfosec.com/rd/arm/?utm_source=rss&utm_medium=rss&utm_campaign=arm Mon, 28 Nov 2022 17:10:31 +0000 https://www.ainfosec.com/?page_id=16643 ...]]>

R&D  >  ARM

RESEARCH & DEVELOPMENT

ARM: Anti-Ransomware Machine

Signature-less Ransomware Defense for Endpoints 

OVERVIEW

ARM provides simple but powerful protection against all kinds of ransomware and eliminates the need for large, expensive defensive software suites.

KEY FEATURES
  • No signatures to update
  • Single independent application
  • Simple to use and configure
  • Cost effective

Reach Out Today

Push your mission forward with the right partner.
]]>
Adversarial Testing https://www.ainfosec.com/rd/adversarial-testing?utm_source=rss&utm_medium=rss&utm_campaign=adversarial-testing Mon, 28 Nov 2022 15:24:48 +0000 https://www.ainfosec.com/?page_id=16607 ...]]>

R&D  >  Adversarial Testing

RESEARCH & DEVELOPMENT

Adversarial Testing

Achieving Results Using an Alternate Perspective 

OVERVIEW

In a world where everything is a target, AIS has expertise in analyzing hardware and software systems from an adversarial perspective. This type of approach has proven results in identifying vulnerabilities and providing mitigation techniques. Let us find the vulnerabilities, so your adversaries don’t.

KEY FEATURES
  • Reverse Engineering
  • Software Characterization
  • Functional Testing
  • Hardware and Software Security

Reach Out Today

Push your mission forward with the right partner.
]]>
ABED https://www.ainfosec.com/rd/abed/?utm_source=rss&utm_medium=rss&utm_campaign=abed Tue, 22 Nov 2022 19:22:59 +0000 https://www.ainfosec.com/?page_id=16578 ...]]>

R&D  >  ABED

RESEARCH & DEVELOPMENT

Artificial-Intelligence Based Endpoint Defender (ABED)

Malicious File Detection on Endpoints

OVERVIEW

ABED is a tool designed to autonomously detect and prevent zero-day exploits from infecting your network through file-based attacks that can be deployed, trained and operational in hours. Additionally, ABED has a built-in threat intelligence engine that is capable of identifying concerted attacks against the network in a visual manner that allows even the most novice analyst deeper insight into malicious activity.

KEY FEATURES
  • Based on signatureless, advanced AI/ML algorithms
  • Built-in threat intelligence with visual context of malicious activity
  • False positive rate of < 1%
  • Requires only a small data set to train the model
  • Nominal impact to the network

Reach Out Today

Push your mission forward with the right partner.
]]>