Our company is developing a new application, which has a somewhat large business data object at its core. We decided to try out Entity Framework with code first to abstract the database from the application, but things have gone awry. The business object is composed of approximately 60 classes, and in total around 600 properties; however, it is a tree structure and no crossover/backtracking pointers are present.
Our test was to add a single, uninitialized instance of the class to the database. Using DbContext.Add on our data structure took 8 minutes on my development machine. Is this the expected performance of an object this size? Is there a list of common problems that causes poor performance with Entity Framework? I feel I need some help with this.
Some more data points: There are 27 elements in the first level under the root of the business object. With 3 elements present (the rest commented out), the time to add is 4.5 seconds. With 5 elements present, it is 11.8 seconds. With 8 elements present, it is 1 minute 12.5 seconds. Obviously, the size of these elements varies significantly, but these does seem to indicate a systematic problem of some sort.
...Our test was to add a single, uninitialized instance of the class to the database. Using DbContext.Add...
Did you make sure that your Code-First model is created and loaded into memory before you've called Add
? I mean the following: If you use a test code like this ...
using (var context = new MyContext())
{
var myHugeBusinessObject = CreateItSomeHow();
context.HugeBusinessObjects.Add(myHugeBusinessObject);
context.SaveChanges();
}
... and it's the first time that you are using the context in your test application Add
will actually spent some time to build the EF model in memory before it starts to add the object to the context.
You can separate these two steps simply by adding a dummy method before calling Add
, for instance something like:
context.HugeBusinessObjects.Count();
I've built a test:
public class MyClass
{
public int Id { get; set; }
public string P1 { get; set; }
// ... P2 to P49
public string P50 { get; set; }
public MyClass Child1 { get; set; }
// ... Child1 to Child26
public MyClass Child27 { get; set; }
}
With this I created an object:
var my = new MyClass();
MyClass child = my;
for (int i = 0; i < 100; i++)
{
child.Child1 = new MyClass();
child = child.Child1;
}
child = my;
for (int i = 0; i < 100; i++)
{
child.Child2 = new MyClass();
child = child.Child2;
}
// and so on up to Child27
So this object graph has 2700 child objects with 50 scalar properties each. Then I tested this code:
using (var context = new MyContext())
{
var my = CreateWithTheCodeAbove();
context.MyClassSet.Count();
context.MyClassSet.Add(my);
context.SaveChanges();
}
...Count()
(= building the EF model) needs roughly 25 seconds. Add
needs 1 second. (Changing 100 to 1000 in the loops above (having then 27000 objects in the graph) increases the time for Add
almost linearly to 9-10 seconds.) This result is independent of setting AutoDetectChangesEnabled
to true
or false
.)
The next interesting result: If I add 20 navigation properties more (Child28
to Child47
) to MyClass
the time spent with building the model (Count()
in the example code) explodes to 140 seconds. The duration of Add
only increases linearly with the additional properties.
So, my hypothesis is: You are not actually measuring the time to add your business object to the context but the time EF needs to build the EF model in memory. The time to build the model seems to grow exponentially with the complexity of the model: Number of navigation properties on a class and perhaps also the number of different involved classes.
To separate these steps test with some dummy call like suggested above. If you already have this separation... omg, forget this post.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With