Click here to Skip to main content
Click here to Skip to main content
Articles » Database » NoSQL » General » Downloads
 
Add your own
alternative version

RaptorDB - the Key Value Store

, 22 Jan 2012 CPOL
Smallest, fastest embedded nosql persisted dictionary using b+tree or MurMur hash indexing. (Now with Hybrid WAH bitmap indexes)
RaptorDB_v1.0.zip
RaptorDB
testing
RaptorDB_v1.1.zip
Indexes
Storage
Views
RaptorDB_v1.2.zip
RaptorDB_v1.3.zip
RaptorDB_v1.4.1.zip
Properties
RaptorDB_v1.4.zip
RaptorDB_v1.5.1.zip
RaptorDB_v1.5.2.zip
RaptorDB_v1.5.zip
RaptorDB_v1.6.zip
RaptorDB_v1.7.5.zip
UnitTests
Properties
UnitTests.nunit
RaptorDB_v1.7.zip
_svn
all-wcprops
dir-prop-base
entries
prop-base
props
text-base
AssemblyInfo.cs.svn-base
tmp
prop-base
props
text-base
UnitTests.nunit
RaptorDB_v1.8.zip
DataTypes
UnitTests release.nunit
UnitTests.nunit
UnitTests.nunit
using System;
using System.Collections.Generic;
using System.Linq;
using System.Text;
using NUnit.Framework;
using RaptorDB;
using System.IO;

namespace UnitTests
{
    [TestFixture]
    public class Tests
    {
        static string string1kb =
            // ~1kb
            "{\"$type\":\"BizFX.TestApp.Entity.Customer, BizFX.TestApp.Entity, Version=1.0.0.0, Culture=neutral, PublicKeyToken=426204062733118a\",\"Address\":null,\"Code\":null,\"Phone\":null,\"email\":null,\"Mobile\":null,\"ContactName\":null,\"Comments\":null,\"GUID\":\"iNIaPond7k6McmSStz14kA==\",\"BaseInfo\":{\"$type\":\"BizFX.Entity.BaseInfo, BizFX.Entity, Version=2.0.0.0, Culture=neutral, PublicKeyToken=e5d192f5e46064af\",\"RevisionNumber\":0,\"CreateDate\":\"2011-04-06 10:16:50\",\"SkipSync\":false,\"SkipDocs\":false,\"SkipRunning\":false,\"DeleteRevisions\":false,\"AssemblyFilename\":\"BizFX.TestApp.Entity.Customer, BizFX.TestApp.Entity, Version=1.0.0.0, Culture=neutral, PublicKeyToken=426204062733118a\",\"TypeName\":\"BizFX.TestApp.Entity.Customer\"},\"SecurityInfo\":{\"$type\":\"BizFX.Entity.SecurityInfo, BizFX.Entity, Version=2.0.0.0, Culture=neutral, PublicKeyToken=e5d192f5e46064af\",\"WinUserName\":\"\",\"AppUserName\":\"\",\"GUID\":\"FAfCsxxJOUuLJITZj005Ow==\",\"LoginName\":\"\",\"UserName\":\"\",\"MachineName\":\"\",\"UserDomainName\":\"\"},\"Description\":\"Base entity description.\",\"Name\":\"BaseEntity\"}";

        static string string100b =
            // ~100 bytes
            "{\"$type\":\"BizFX.TestApp.Entity.Customer, BizFX.TestApp.Entity, Version=1.0.0.0, Culture=neutral, PublicKeyToken=426204062733118a\"";

        [Test]
        public void String_As_Key_btree()
        {
            string_as_key(INDEXTYPE.BTREE);
        }


        [Test]
        public void String_As_Key_hash()
        {
            string_as_key(INDEXTYPE.HASH);
        }

        private static void string_as_key(INDEXTYPE type)
        {
            int count = 40000;
            Console.WriteLine("Count = " + count);
            RaptorDB.RaptorDB rap = RaptorDB.RaptorDB.Open("docs\\strings" + type, 30, true, type);
            rap.IndexingTimerSeconds = 1;
            rap.InMemoryIndex = true;
            string key = "some very long string";

            for (int i = 0; i < count; i++)
            {
                string ss = key + i.ToString("000000");
                rap.Set(key, ss);
            }
            //System.Threading.Thread.Sleep(5000);
            rap.SaveIndex(true);
            int j = 0;
            for (int i = 0; i < count; i++)
            {
                string ss = key + i.ToString("000000");
                byte[] bb = null;
                if (rap.Get(key, out bb) == false)
                    j++;// Console.WriteLine("error");
            }
            Console.WriteLine("Error count = " + j);
            Assert.AreEqual(j, 0);
        }

        [Test]
        public void Duplicate_Guid_Key_btree()
        {
            duplicate_guid_key(INDEXTYPE.BTREE, true);
        }


        [Test]
        public void Duplicate_Guid_Key_hash()
        {
            duplicate_guid_key(INDEXTYPE.HASH, true);
        }

        private static void duplicate_guid_key(INDEXTYPE type, bool allowdups)
        {
            byte[] bb = System.Text.Encoding.UTF8.GetBytes(string1kb);
            RaptorDB.RaptorDB rap = RaptorDB.RaptorDB.Open("docs\\duplicates" + type, 16, allowdups, type);
            rap.IndexingTimerSeconds = 1;
            rap.InMemoryIndex = true;
            Guid g = Guid.NewGuid();
            Console.WriteLine("saving...");
            rap.Set(Guid.NewGuid(), bb);
            for (int i = 0; i < 20000; i++)
            {
                rap.Set(g, bb);
            }
            rap.SaveIndex(true);
            foreach (var l in rap.GetDuplicates(g.ToByteArray()))
            {
                byte[] dup = rap.FetchDuplicate(l);
                int i = dup.Length;
                Assert.AreEqual(i, bb.Length);
            }

            //foreach (var b in rap.EnumerateStorageFile())
            //{
            //    g = new Guid(b.Key);
            //    string sss = System.Text.Encoding.UTF8.GetString(b.Value);
            //    Console.Write(g);
            //}
        }


        [Test]
        public static void Dave_Killer_Test()
        {
            var db = RaptorDB.RaptorDB.Open("c:\\RaptorDbTest\\RawFileOne", 16, true, INDEXTYPE.BTREE);
            db.InMemoryIndex = true;
            //db.IndexingTimerSeconds = 1000;
            var guids = new List<Guid>();
            var sw = new System.Diagnostics.Stopwatch();
            const int trials = 1000000;
            for (int x = 0; x < trials; x++)
            {
                guids.Add(Guid.NewGuid());
            }
            sw.Start();
            for (int x = 0; x < trials; x++)
            {
                var id = guids[x];
                var key = id.ToByteArray();
                var value = id.ToString();
                db.Set(id, Encoding.UTF8.GetBytes(value));
            }
            sw.Stop();
            Console.Out.WriteLine("\n\n\nWriting {0} items took: {1}", trials, sw.Elapsed);
            var writeSpeed = trials * 1000.0d / sw.ElapsedMilliseconds;
            Console.Out.WriteLine("\nAverage write speed: {0} items/second.", writeSpeed);
            Console.Out.Write("Saving indexes...");
            sw.Reset();
            sw.Start();
            db.SaveIndex(true);
            sw.Stop();
            Console.Out.WriteLine("\n\nSaving indices took {0} ms", sw.ElapsedMilliseconds);
            sw.Reset(); sw.Start();
            int readCount = 0;
            byte[] valBytes;
            for (int x = 0; x < trials; x++)
            {
                var id = guids[x];

                if (db.Get(id, out valBytes))
                {
                    readCount++;
                }
            }
            sw.Stop();
            Console.Out.WriteLine("\n\n\nSuccessfully read {0} of {1} items took: {2}...", readCount, trials, sw.Elapsed);
            var readSpeed = readCount * 1000.0d / sw.ElapsedMilliseconds;
            Console.Out.WriteLine("\nAverage index hit/read speed: {0:0.##} items/second.", readSpeed);
            Console.WriteLine("\n\n\nAdding new values at the same keys (Duplicates enabled)...");
            sw.Reset();
            var newValues = new List<string>(trials);
            for (int x = 0; x < trials; x++)
            {
                newValues.Add(guids[x] + "ABC");
            }
            sw.Start();
            for (int x = 0; x < trials; x++)
            {
                db.Set(
                    //Guid.NewGuid()
                     guids[x]
                    , Encoding.UTF8.GetBytes(newValues[x]));
            }
            sw.Stop();
            Console.Out.WriteLine("Set {0} duplicate keys in {1}", trials, sw.Elapsed);
            var setDuplicateSpeed = trials * 1000.0d / sw.ElapsedMilliseconds;
            Console.Out.WriteLine("Average 'set duplicate key value speed' = {0} items/second.", setDuplicateSpeed);
            /* Again I MUST wait for the indexing to finish or I'm toast */
            sw.Reset(); sw.Start();

            db.SaveIndex(true);
            sw.Stop();
            Console.Out.WriteLine("\n\nSaving indices took {0} ms", sw.ElapsedMilliseconds);

            sw.Reset();
            sw.Start();
            Console.Out.WriteLine("\nEnumerating sets of 2 duplicates...");
            int duplicateCount = 0;
            for (int x = 0; x < trials; x++)
            {
                //try
                {
                    var ints = db.GetDuplicates(guids[x].ToByteArray());
                    duplicateCount += ints.Count();
                }
                //catch (Exception ex)
                //{
                //    Console.WriteLine("" + ex);
                //}
            }
            sw.Stop();
            Console.WriteLine("\n\n\nSuccessfully enumerated {0} items and found {1} duplicates in {2}", trials, duplicateCount, sw.Elapsed);
            var duplicateSpeed = duplicateCount * 1000.0d / sw.ElapsedMilliseconds;
            Console.WriteLine("\nSpeed of enumerating duplicates: {0} items/second.", duplicateSpeed);
        }

        [Test]
        public static void SafeDictionary_Test()
        {
            int count = 500000;
            Console.WriteLine("count = " + count);
            SafeDictionary<byte[], int> d = new SafeDictionary<byte[], int>(20, new ByteArrayComparer());
            List<Guid> guids = new List<Guid>();
            for (int i = 0; i < count; i++)
            {
                guids.Add(Guid.NewGuid());
            }
            for (int i = 0; i < 10; i++)
                d.Add(guids[i].ToByteArray(), i);

            for (int i = 0; i < 10; i++)
            {
                int j = -1;
                bool b = d.TryGetValue(guids[i].ToByteArray(), out j);
                Assert.AreEqual(i, j);
                if (i != j)
                {
                    Console.Write("x" + j);
                }
            }
        }

        [Test]
        public static void Duplicates_Set_and_Fetch_btree()
        {
            var db = RaptorDB.RaptorDB.Open("c:\\RaptorDbTest\\duptestfetch", 16, true, INDEXTYPE.BTREE);
            db.InMemoryIndex = true;
            //db.IndexingTimerSeconds = 1000;
            int guidcount = 1000;
            int dupcount = 100;
            var guids = new List<Guid>();
            for (int i = 0; i < guidcount; i++)
                guids.Add(Guid.NewGuid());

            foreach (Guid g in guids)
            {
                for (int i = 0; i < dupcount; i++)
                {
                    string s = "" + g + " " + i;
                    db.Set(g, Encoding.UTF8.GetBytes(s));
                }
            }
            db.SaveIndex(true);

            foreach (Guid g in guids)
            {
                int j = 0;
                foreach (int i in db.GetDuplicates(g.ToByteArray()))
                {
                    byte[] b = db.FetchDuplicate(i);
                    string s = Encoding.UTF8.GetString(b);
                    if (s.StartsWith(g.ToString()) == false)
                        Console.WriteLine("guid not correct = " + g + " returned = " + s);
                    else
                        j++;
                }
                if (j < dupcount - 1)
                {
                    Console.WriteLine("" + g + " count = " + j);
                    Assert.Fail();
                }
                else
                    Console.WriteLine("" + g + " = OK");
            }

            db.Shutdown();
        }

        [Test]
        public static void One_Million_Set_Get_BTREE()
        {
            One_Million_Set_Get(INDEXTYPE.BTREE, 1000000, true);
        }

        [Test]
        public static void Twenty_Million_Set_Get_BTREE()
        {
            Console.WriteLine("Twenty million insert test");
            Console.WriteLine("This test will use a peak of 1.6Gb ram for guid keys");
            Console.WriteLine("This test will run for about 16 minutes depending on your hardware");
            One_Million_Set_Get(INDEXTYPE.BTREE, 20 * 1000000, true);
        }

        [Test]
        public static void One_Million_Set_Get_HASH()
        {
            One_Million_Set_Get(INDEXTYPE.HASH, 1000000, true);
        }

        private static void One_Million_Set_Get(INDEXTYPE type, int count, bool inmem)
        {
            Console.WriteLine("One million test on " + type);
            var db = RaptorDB.RaptorDB.Open("c:\\RaptorDbTest\\1million" + type, 16, false, type);
            db.InMemoryIndex = inmem;
            //db.IndexingTimerSeconds = 1000;
            Console.Write("Building guid list...");
            var guids = new List<Guid>();
            for (int i = 0; i < count; i++)
                guids.Add(Guid.NewGuid());
            Console.WriteLine("done");
            DateTime dt = DateTime.Now;
            int c = 0;
            foreach (Guid g in guids)
            {
                string s = "" + g;
                db.Set(g, Encoding.Unicode.GetBytes(s));
                c++;
                if (c % 10000 == 0)
                    Console.Write(".");
                if (c % 100000 == 0)
                    Console.WriteLine("time = " + DateTime.Now.Subtract(dt).TotalSeconds);
            }
            Console.WriteLine("Flushing index...");
            db.SaveIndex(true);
            Console.WriteLine(count.ToString("#,#") + " save total time = " + DateTime.Now.Subtract(dt).TotalSeconds);
            dt = DateTime.Now;
            foreach (Guid g in guids)
            {
                byte[] val;
                if (db.Get(g, out val))
                {
                    string s = Encoding.Unicode.GetString(val);
                    if (s.Equals("" + g) == false)
                        Assert.Fail("data does not match " + g);
                }
                else
                    Assert.Fail("item not found " + g);
            }
            Console.WriteLine("fetch total time = " + DateTime.Now.Subtract(dt).TotalSeconds);
            Console.WriteLine("ALL DONE OK");
            db.Shutdown();
        }

        [Test]
        public static void WAH_Bitarray_test()
        {
            WAHBitArray ba = new WAHBitArray();

            Random r = new Random();

            // ----------------------------------------------------------
            Console.Write("Random bits set...");
            for (int i = 0; i < 100000; i++)
            {
                if (r.Next(200) == 0)
                {
                    ba.Set(i, true);
                }
            }
            uint[] ui = ba.GetCompressed();
            WAHBitArray b2 = new WAHBitArray(ba.UsingIndexes ? WAHBitArray.TYPE.Indexes : WAHBitArray.TYPE.Compressed_WAH, ui);

            WAHBitArray result = b2.Xor(ba);
            Assert.AreEqual(0, result.CountOnes());
            Console.WriteLine(" done");



            // ----------------------------------------------------------
            Console.Write("All 1 set...");
            ba = new WAHBitArray();
            for (int i = 0; i < 100000; i++)
                ba.Set(i, true);
            ui = ba.GetCompressed();
            b2 = new WAHBitArray(ba.UsingIndexes ? WAHBitArray.TYPE.Indexes : WAHBitArray.TYPE.Compressed_WAH, ui);

            result = b2.Xor(ba);
            Assert.AreEqual(0, result.CountOnes());
            Console.WriteLine(" done");



            // ----------------------------------------------------------
            Console.Write("Alternate 1 set...");
            ba = new WAHBitArray();
            for (int i = 0; i < 100000; i++)
            {
                if (i % 2 == 0)
                    ba.Set(i, true);
            }
            ui = ba.GetCompressed();
            b2 = new WAHBitArray(ba.UsingIndexes ? WAHBitArray.TYPE.Indexes : WAHBitArray.TYPE.Compressed_WAH, ui);

            result = b2.Xor(ba);
            Assert.AreEqual(0, result.CountOnes());
            Console.WriteLine(" done");



            // ----------------------------------------------------------
            Console.Write("Alternate 1 xor not self set...");
            ba = new WAHBitArray();
            for (int i = 0; i < 100000; i++)
            {
                if (i % 2 == 0)
                    ba.Set(i, true);
            }
            ui = ba.GetCompressed();
            b2 = new WAHBitArray(ba.UsingIndexes ? WAHBitArray.TYPE.Indexes : WAHBitArray.TYPE.Compressed_WAH, ui);

            result = b2.Xor(ba.Not());
            Assert.AreEqual(100000, result.CountOnes());
            Console.WriteLine(" done");



            // ----------------------------------------------------------
            Console.Write("half 1 set...");
            ba = new WAHBitArray();
            for (int i = 50000; i < 100000; i++)
            {
                ba.Set(i, true);
            }
            ui = ba.GetCompressed();
            b2 = new WAHBitArray(ba.UsingIndexes ? WAHBitArray.TYPE.Indexes : WAHBitArray.TYPE.Compressed_WAH, ui);

            result = b2.Xor(ba);
            Assert.AreEqual(0, result.CountOnes());
            Console.WriteLine(" done");



            // ----------------------------------------------------------
            Console.Write("other half 1 set...");
            ba = new WAHBitArray();
            for (int i = 0; i < 50000; i++)
            {
                ba.Set(i, true);
            }
            ba.Set(100000, true);
            ui = ba.GetCompressed();
            b2 = new WAHBitArray(ba.UsingIndexes ? WAHBitArray.TYPE.Indexes : WAHBitArray.TYPE.Compressed_WAH, ui);

            result = b2.Xor(ba);
            Assert.AreEqual(0, result.CountOnes());
            Console.WriteLine(" done");
        }

        [Test]
        public static void WAH_Multiple_duplicates_test()
        {
            for (int i = 0; i < 100; i++)
            {
                Console.WriteLine("i = " + i);
                WAHBitArray ba = new WAHBitArray();
                for (int j = 0; j < 1000; j++)
                {
                    ba.Set(i * 1000 + j, true);
                }
                uint[] ui = ba.GetCompressed();
                WAHBitArray b2 = new WAHBitArray(ba.UsingIndexes ? WAHBitArray.TYPE.Indexes : WAHBitArray.TYPE.Compressed_WAH, ui);

                WAHBitArray result = b2.Xor(ba);
                Assert.AreEqual(0, result.CountOnes());
                Console.WriteLine(" done");
            }
        }
    }
}

By viewing downloads associated with this article you agree to the Terms of Service and the article's licence.

If a file you wish to view isn't highlighted, and is a text file (not binary), please let us know and we'll add colourisation support for it.

License

This article, along with any associated source code and files, is licensed under The Code Project Open License (CPOL)

Share

About the Author

Mehdi Gholam
Architect
United Kingdom United Kingdom
Mehdi first started programming when he was 8 on BBC+128k machine in 6512 processor language, after various hardware and software changes he eventually came across .net and c# which he has been using since v1.0.
He is formally educated as a system analyst Industrial engineer, but his programming passion continues.
 
* Mehdi is the 5th person to get 6 out of 7 Platinums on CodeProject (13th Jan'12)

| Advertise | Privacy | Terms of Use | Mobile
Web03 | 2.8.1411023.1 | Last Updated 23 Jan 2012
Article Copyright 2011 by Mehdi Gholam
Everything else Copyright © CodeProject, 1999-2014
Layout: fixed | fluid