ZipPy: Fast method to classify text as AI or human-generated
 
 
 
 
Go to file
Sara Tavares 294ad04638
Update README.md
2023-11-07 21:28:11 +00:00
.github/workflows Update pytest.yml 2023-09-27 16:54:43 -06:00
inch Update browser extension repo to v0.3.1 2023-09-21 13:33:36 -06:00
nlzmadetect Fix ai-generated.txt path for Nim tests 2023-10-27 11:56:32 -06:00
samples
test_results Update plot_rocs to point to new test result directory 2023-10-15 09:52:43 -06:00
zippy Ensure the ai-generated.txt is included in the built package 2023-10-27 17:21:58 +00:00
.gitignore
.gitmodules
LICENSE
README.md Update README.md 2023-11-07 21:28:11 +00:00
ai_detect_roc.png Completed evaluation of contentatscale.ai and added zlib support to both the Python and Nim/JS implementations 2023-09-26 07:51:41 -06:00
burstiness.py
contentatscale_detect.py Initial results from contentatscale.ai added, and removed the OpenAI detector from the ROC chart plotter 2023-09-21 16:37:28 -06:00
crossplag_detect.py
gptzero_detect.py
openai_detect.py
plot_rocs.py Update plot_rocs to point to new test result directory 2023-10-15 09:52:43 -06:00
preset_plot_rocs.py Added preset analysis 2023-10-15 04:42:53 -06:00
requirements.txt Add Windows support and a requirements.txt 2023-10-27 09:54:13 -06:00
roberta_detect.py
roberta_local.py
setup.py Ensure the ai-generated.txt is included in the built package 2023-10-27 17:21:58 +00:00
test_contentatscale_detect.py Completed evaluation of contentatscale.ai and added zlib support to both the Python and Nim/JS implementations 2023-09-26 07:51:41 -06:00
test_crossplag_detect.py
test_gptzero_detect.py
test_openai_detect.py
test_roberta_detect.py
test_zippy_detect.py Ensure the ai-generated.txt is included in the built package 2023-10-27 17:21:58 +00:00

README.md

ZipPy: Fast method to classify text as AI or human-generated

This is a research repo for fast AI detection using compression. While there are a number of existing LLM detection systems, they all use a large model trained on either an LLM or its training data to calculate the probability of each word given the preceding, then calculate a score where the more high-probability tokens are more likely to be AI-originated. Techniques and tools in this repo are looking for faster approximation to be embeddable and more scalable.

Compression-based detector (zippy.py and nlzmadetect)

ZipPy uses either the LZMA or zlib compression ratios as a way to indirectly measure the perplexity of a text. Compression ratios have been used in the past to detect anomalies in network data for intrusion detection, so if perplexity is roughly a measure of anomalous tokens, it may be possible to use compression to detect low-perplexity text. LZMA and zlib create a dictionary of seen tokens and then use though in place of future tokens. The dictionary size, token length, etc. are all dynamic (though influenced by the 'preset' of 0-9--with 0 being the fastest but worse compression than 9). The basic idea is to 'seed' a compression stream with a corpus of AI-generated text (ai-generated.txt) and then measure the compression ratio of just the seed data with that of the sample appended. Samples that follow more closely in word choice, structure, etc. will achieve a higher compression ratio due to the prevalence of similar tokens in the dictionary, novel words, structures, etc. will appear anomalous to the seeded dictionary, resulting in a worse compression ratio.

Current evaluation

Some of the leading LLM detection tools are: OpenAI's model detector (v2), Content at Scale, GPTZero, CrossPlag's AI detector, and Roberta. Here are each of them compared with both the LZMA and zlib detector across the test datasets:

ROC curve of detection tools

Usage

ZipPy will read files passed as command-line arguments or will read from stdin to allow for piping of text to it.

First, build and install the tool:

$ python3 setup.py build && python3 setup.py install

It will install a new script (zippy) that you can use directly:

$ zippy -h
usage: zippy [-h] [-p P] [-e {zlib,lzma,brotli,ensemble}] [-s | sample_files ...]

positional arguments:
  sample_files          Text file(s) containing the sample to classify

options:
  -h, --help            show this help message and exit
  -p P                  Preset to use with compressor, higher values are slower but provide better compression
  -e {zlib,lzma,brotli,ensemble}
                        Which compression engine to use: lzma, zlib, brotli, or an ensemble of all engines
  -s                    Read from stdin until EOF is reached instead of from a file
$ zippy samples/human-generated/about_me.txt 
samples/human-generated/about_me.txt
('Human', 0.06013429262166636)

If you want to use the ZipPy technology in your browser, check out the Chrome extension or the Firefox extension that runs ZipPy in-browser to flag potentially AI-generated content.