Merge pull request #166 from yankejustin/UnitTests

Added another Unit Test
This commit is contained in:
MaxXor 2015-05-23 09:51:06 +02:00
commit 8e203d0672
5 changed files with 267 additions and 23 deletions

View File

@ -55,9 +55,9 @@
</Choose>
<ItemGroup>
<Compile Include="Core\Compression\JpgCompression.Tests.cs" />
<Compile Include="Core\Compression\SafeQuickLZ.Tests.cs" />
<Compile Include="Core\Encryption\AES.Tests.cs" />
<Compile Include="Core\Encryption\SHA256.Tests.cs" />
<Compile Include="Core\Information\GeoIP.Tests.cs" />
<Compile Include="Properties\AssemblyInfo.cs" />
</ItemGroup>
<ItemGroup>
@ -66,7 +66,9 @@
<Name>Client</Name>
</ProjectReference>
</ItemGroup>
<ItemGroup />
<ItemGroup>
<Folder Include="Core\Information\" />
</ItemGroup>
<Choose>
<When Condition="'$(VisualStudioVersion)' == '10.0' And '$(IsCodedUITest)' == 'True'">
<ItemGroup>

View File

@ -0,0 +1,131 @@
using System;
using Microsoft.VisualStudio.TestTools.UnitTesting;
using xClient.Core.Compression;
namespace xClient.Tests.Core.Compression
{
[TestClass]
public class SafeQuickLZTests
{
// Tests using pseudo-randomly generated data.
#region Random Data
/*
* Purpose: To validate a small amount of data after compression/decompression
* using SafeQuickLZ with level 1 compression.
*/
[TestMethod]
[TestCategory("Compression")]
public void SmallDataTestLevel1()
{
SafeQuickLZ safeQuickLZtest = new SafeQuickLZ();
byte[] SmallData = new byte[100];
// Fill the small data array with random data.
new Random().NextBytes(SmallData);
// Store the compressed data.
byte[] SmallDataCompressed = safeQuickLZtest.Compress(SmallData, 0, SmallData.Length, 1);
// The original should not equal the compressed data.
Assert.AreNotEqual(SmallData, SmallDataCompressed, "Original data is equal to the compressed data!");
// Store the decompressed data.
byte[] SmallDataDecompressed = safeQuickLZtest.Decompress(SmallDataCompressed, 0, SmallDataCompressed.Length);
// The compressed data should not equal the decompressed data.
Assert.AreNotEqual(SmallDataCompressed, SmallDataDecompressed, "Compressed data is equal to the decompressed data!");
// The original data must equal the decompressed data; must be able to make a round-trip.
CollectionAssert.AreEqual(SmallData, SmallDataDecompressed, "Original data does not match the decompressed data!");
}
/*
* Purpose: To validate a small amount of data after compression/decompression
* using SafeQuickLZ with level 3 compression.
*/
[TestMethod]
[TestCategory("Compression")]
public void SmallDataTestLevel3()
{
SafeQuickLZ safeQuickLZtest = new SafeQuickLZ();
byte[] SmallData = new byte[100];
// Fill the small data array with random data.
new Random().NextBytes(SmallData);
// Store the compressed data.
byte[] SmallDataCompressed = safeQuickLZtest.Compress(SmallData, 0, SmallData.Length, 3);
// The original should not equal the compressed data.
Assert.AreNotEqual(SmallData, SmallDataCompressed, "Original data is equal to the compressed data!");
// Store the decompressed data.
byte[] SmallDataDecompressed = safeQuickLZtest.Decompress(SmallDataCompressed, 0, SmallDataCompressed.Length);
// The compressed data should not equal the decompressed data.
Assert.AreNotEqual(SmallDataCompressed, SmallDataDecompressed, "Compressed data is equal to the decompressed data!");
// The original data must equal the decompressed data; must be able to make a round-trip.
CollectionAssert.AreEqual(SmallData, SmallDataDecompressed, "Original data does not match the decompressed data!");
}
/*
* Purpose: To validate a large amount of data after compression/decompression
* using SafeQuickLZ with level 1 compression.
*/
[TestMethod]
[TestCategory("Compression")]
public void BigDataTestLevel1()
{
SafeQuickLZ safeQuickLZtest = new SafeQuickLZ();
byte[] BigData = new byte[100000];
// Fill the big data array with random data.
new Random().NextBytes(BigData);
// Store the compressed data.
byte[] BigDataCompressed = safeQuickLZtest.Compress(BigData, 0, BigData.Length, 1);
// The original should not equal the compressed data.
Assert.AreNotEqual(BigData, BigDataCompressed, "Original data is equal to the compressed data!");
// Store the decompressed data.
byte[] BigDataDecompressed = safeQuickLZtest.Decompress(BigDataCompressed, 0, BigDataCompressed.Length);
// The compressed data should not equal the decompressed data.
Assert.AreNotEqual(BigDataCompressed, BigDataDecompressed, "Compressed data is equal to the decompressed data!");
// The original data must equal the decompressed data; must be able to make a round-trip.
CollectionAssert.AreEqual(BigData, BigDataDecompressed, "Original data does not match the decompressed data!");
}
/*
* Purpose: To validate a large amount of data after compression/decompression
* using SafeQuickLZ with level 3 compression.
*/
[TestMethod]
[TestCategory("Compression")]
public void BigDataTestLevel3()
{
SafeQuickLZ safeQuickLZtest = new SafeQuickLZ();
byte[] BigData = new byte[100000];
// Fill the big data array with random data.
new Random().NextBytes(BigData);
// Store the compressed data.
byte[] BigDataCompressed = safeQuickLZtest.Compress(BigData, 0, BigData.Length, 3);
// The original should not equal the compressed data.
Assert.AreNotEqual(BigData, BigDataCompressed, "Original data is equal to the compressed data!");
// Store the decompressed data.
byte[] BigDataDecompressed = safeQuickLZtest.Decompress(BigDataCompressed, 0, BigDataCompressed.Length);
// The compressed data should not equal the decompressed data.
Assert.AreNotEqual(BigDataCompressed, BigDataDecompressed, "Compressed data is equal to the decompressed data!");
// The original data must equal the decompressed data; must be able to make a round-trip.
CollectionAssert.AreEqual(BigData, BigDataDecompressed, "Original data does not match the decompressed data!");
}
#endregion
}
}

View File

@ -1,21 +0,0 @@
using Microsoft.VisualStudio.TestTools.UnitTesting;
using xClient.Core.Information;
namespace xClient.Tests.Core.Information
{
[TestClass]
public class GeoIPTests
{
[TestMethod]
public void GetGeoIPTest()
{
var ipInformation = new GeoIP();
Assert.IsNotNull(ipInformation.City);
Assert.IsNotNull(ipInformation.Country);
Assert.IsNotNull(ipInformation.CountryCode);
Assert.IsNotNull(ipInformation.Region);
Assert.IsNotNull(ipInformation.WanIp);
}
}
}

View File

@ -0,0 +1,131 @@
using System;
using Microsoft.VisualStudio.TestTools.UnitTesting;
using xServer.Core.Compression;
namespace xServer.Tests.Core.Compression
{
[TestClass]
public class SafeQuickLZTests
{
// Tests using pseudo-randomly generated data.
#region Random Data
/*
* Purpose: To validate a small amount of data after compression/decompression
* using SafeQuickLZ with level 1 compression.
*/
[TestMethod]
[TestCategory("Compression")]
public void SmallDataTestLevel1()
{
SafeQuickLZ safeQuickLZtest = new SafeQuickLZ();
byte[] SmallData = new byte[100];
// Fill the small data array with random data.
new Random().NextBytes(SmallData);
// Store the compressed data.
byte[] SmallDataCompressed = safeQuickLZtest.Compress(SmallData, 0, SmallData.Length, 1);
// The original should not equal the compressed data.
Assert.AreNotEqual(SmallData, SmallDataCompressed, "Original data is equal to the compressed data!");
// Store the decompressed data.
byte[] SmallDataDecompressed = safeQuickLZtest.Decompress(SmallDataCompressed, 0, SmallDataCompressed.Length);
// The compressed data should not equal the decompressed data.
Assert.AreNotEqual(SmallDataCompressed, SmallDataDecompressed, "Compressed data is equal to the decompressed data!");
// The original data must equal the decompressed data; must be able to make a round-trip.
CollectionAssert.AreEqual(SmallData, SmallDataDecompressed, "Original data does not match the decompressed data!");
}
/*
* Purpose: To validate a small amount of data after compression/decompression
* using SafeQuickLZ with level 3 compression.
*/
[TestMethod]
[TestCategory("Compression")]
public void SmallDataTestLevel3()
{
SafeQuickLZ safeQuickLZtest = new SafeQuickLZ();
byte[] SmallData = new byte[100];
// Fill the small data array with random data.
new Random().NextBytes(SmallData);
// Store the compressed data.
byte[] SmallDataCompressed = safeQuickLZtest.Compress(SmallData, 0, SmallData.Length, 3);
// The original should not equal the compressed data.
Assert.AreNotEqual(SmallData, SmallDataCompressed, "Original data is equal to the compressed data!");
// Store the decompressed data.
byte[] SmallDataDecompressed = safeQuickLZtest.Decompress(SmallDataCompressed, 0, SmallDataCompressed.Length);
// The compressed data should not equal the decompressed data.
Assert.AreNotEqual(SmallDataCompressed, SmallDataDecompressed, "Compressed data is equal to the decompressed data!");
// The original data must equal the decompressed data; must be able to make a round-trip.
CollectionAssert.AreEqual(SmallData, SmallDataDecompressed, "Original data does not match the decompressed data!");
}
/*
* Purpose: To validate a large amount of data after compression/decompression
* using SafeQuickLZ with level 1 compression.
*/
[TestMethod]
[TestCategory("Compression")]
public void BigDataTestLevel1()
{
SafeQuickLZ safeQuickLZtest = new SafeQuickLZ();
byte[] BigData = new byte[100000];
// Fill the big data array with random data.
new Random().NextBytes(BigData);
// Store the compressed data.
byte[] BigDataCompressed = safeQuickLZtest.Compress(BigData, 0, BigData.Length, 1);
// The original should not equal the compressed data.
Assert.AreNotEqual(BigData, BigDataCompressed, "Original data is equal to the compressed data!");
// Store the decompressed data.
byte[] BigDataDecompressed = safeQuickLZtest.Decompress(BigDataCompressed, 0, BigDataCompressed.Length);
// The compressed data should not equal the decompressed data.
Assert.AreNotEqual(BigDataCompressed, BigDataDecompressed, "Compressed data is equal to the decompressed data!");
// The original data must equal the decompressed data; must be able to make a round-trip.
CollectionAssert.AreEqual(BigData, BigDataDecompressed, "Original data does not match the decompressed data!");
}
/*
* Purpose: To validate a large amount of data after compression/decompression
* using SafeQuickLZ with level 3 compression.
*/
[TestMethod]
[TestCategory("Compression")]
public void BigDataTestLevel3()
{
SafeQuickLZ safeQuickLZtest = new SafeQuickLZ();
byte[] BigData = new byte[100000];
// Fill the big data array with random data.
new Random().NextBytes(BigData);
// Store the compressed data.
byte[] BigDataCompressed = safeQuickLZtest.Compress(BigData, 0, BigData.Length, 3);
// The original should not equal the compressed data.
Assert.AreNotEqual(BigData, BigDataCompressed, "Original data is equal to the compressed data!");
// Store the decompressed data.
byte[] BigDataDecompressed = safeQuickLZtest.Decompress(BigDataCompressed, 0, BigDataCompressed.Length);
// The compressed data should not equal the decompressed data.
Assert.AreNotEqual(BigDataCompressed, BigDataDecompressed, "Compressed data is equal to the decompressed data!");
// The original data must equal the decompressed data; must be able to make a round-trip.
CollectionAssert.AreEqual(BigData, BigDataDecompressed, "Original data does not match the decompressed data!");
}
#endregion
}
}

View File

@ -53,6 +53,7 @@
</Otherwise>
</Choose>
<ItemGroup>
<Compile Include="Core\Compression\SafeQuickLZ.Tests.cs" />
<Compile Include="Core\Encryption\AES.Tests.cs" />
<Compile Include="Properties\AssemblyInfo.cs" />
</ItemGroup>