Skip to content

spcl/FBACode

Repository files navigation

Fetch, Build and Analyze Code

The project consists of three major components: fetcher, builder, and analyzer. The first one is responsible for discovering repositories and source codes. The second one downloades the code, attempts build and generates LLVM bitcodes, and sends them to a remote server for storage. Finally, the analyzer downoads the build artifacts from the remote storage server and analyzes them.

We perform the analysis on the AST files produces by the builder. Specifically, the analysis is done by a C++ tool (cxx-langstat) that relies on the Clang compiler infrastructure to parse the ASTs. This introduces a requirement: the langstat tool must be built with the exact same compiler version that the AST files were built with. We achieve this by using docker multi-stage builds.

Requirements

  • Docker: make sure that Docker daemon is running and your user has sufficient permissions to use containers.
  • Python (3.6+) with virtual environments and pip. The following Python packages are installed through venv and used:

Installation

Run the install.py script to create a Python virtual environment fbacode-virtualenv and install dependencies there. Use source fbacode-virtualenv/bin/activate to active the environment.

Use tools/build_docker_images.py to build and update Docker images necessary to run build steps.

Debian installation

Creating the docker containers

First, either pull the build and analyze images from DockerHub using:

docker pull spcleth:debian-bookworm-clang
docker pull spcleth:debian-bookworm-cxxlangstat

Or build the images locally using python3 tools/build_docker_images.py. The order in which to create the docker containers is:

  1. debian-bookworm-clang-base
  2. debian-bookworm-clang-base-beta
  3. debian-bookworm-clang
  4. debian-bookworm-cxxlangstat

When building the debian-bookworm-clang image, make sure the cxx-langstat project can be found as folder inside the root of FBACode.

Fetcher

Fetcher attempts to discover available repositories that could be built. The goal is to search known and popular platforms with open-source code. The output is a JSON file with entries for each project. The current implementation supports GitHub and Debian packages. To enable or disable sources, use the fetch.cfg file and set the active field accordingly

GitHub

The implementation scans GH repositories tagged as C and C++ software and sorts them according to the number of stars.

To avoid running into rate limit issues, please provide your personal access token in the user configuration file.

Debian

The fetcher looks for random projects with C++ code in them, which is found in the debian package API. At the moment only Debian 12 (Bookworm) is supported.

Builder

The builder component is responsible for downloading source code, discovering the build system used for this projects, lookup of existing build instructions and dependencies and attempting a build. For the build, we use LLVM's Clang and generate LLVM IR bitcodes.

The Python implementation uses a thread pool to start multiple build processes at the same time. Each build is invoked in a Docker container for security and compatibility reasons.

Current implementation supports two build systems: CMake projects and Debian source packages.

Try builder with builder.py examples/github-repo.json or builder.py examples/debian.json.

The builder outputs several files to the buildlogs folder:

  • summary-XXX.txt: basically the same output as the console, shows errors and other build statiscics
  • build_details_XXX.json: contains all the information gathered for each project
  • errorstats_XXX.json: A sorted list of all errors found, with packages and numbers of occurences
  • rebuild_XXX.json: A json file with all the failed projects, can be fed to the Builder again
  • dependencies_XXX.json: A sorted list of all missing dependencies found

CMake

Current implementation supports default configuration without any configuration flags.

Make

Current implementation basically runs ./configure and then make.

Autotools

Current implementation basically runs autoreconf, ./configure and then make. Does not work correctly yet.

Travis CI

The builder tries to pick a configuration with os: linux, compiler: clang and arch: amd64 otherwise just picks the first configuration from the build matrix. The apt and snap addons are supported, as well as stages.

Debian Builder

The current implementation uses packages from the Debian 12 (Bookworm, latest LTS at the time) repository. The Packages get downloaded inside the docker container, since we don't know if apt is available on the host system. it basically runs apt-source package, then apt build-dep package to install dependencies and finally dpkg-buildpackage to build it. because the configure and build step is combined into one command, it is not possible to time them separately.

Missing features

  • Currently we don't perform any search for cloned repositories. Furthermore, it's possible that different platforms provide various versions of the same software, e.g. a GitHub project of a program delivered as a Debian package as well.
  • We need a fleet of Docker containers supporting different versions of LLVM.
  • installing dependencies is missing
  • discovering dependencies is not very good yet
  • We should use CI configurations (Travis, CircleCI, GH Actions) and package mangers to discover configuration flags and dependencies.

Analyzer

The analyzer component is responsible for "downloading" the build artifacts from the remote storage server and then to analyze them. It uses a docker image to start a container that will decompress the archive and run cxx-langstat on the AST files.

Example build and analysis on Debian

Here, we will showcase how to build and analyze projects from the Debian repository. We will use the examples/debian-abseil.json file, but it can be replaced with any other debian json database.

Build

python3 builder.py examples/debian-abseil.json

After the build finishes, you should see on the remote server specified inside of build.cfg an archive with the results of the build. Depending on the contents of the config file, the build artifacts might exist on the build machine as well.

Analysis

Make sure the root of FBACode contains a folder named ast_archive. This is where the analyzer will "download" the artifacts from the remote server. Additionally, the build artifacts folder must contain a build_summary.json file. This is similar to a database of the artifacts in that folder. If it does not exist, we recommend using the all_built.json in the build directory used by builder.py (by default it is build/).

python3 analyzer.py <folder containing the build artifacts> # typically of form run_<timestamp>

Then, in the analyze/ folder, there should be separate folders for each project that was analyzed. Each folder contains the statistics for that specific project.

Next, the data_analysis/ folder contains a jupyter notebook that aggregates the per-project statistics to extract high-level insights.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •