Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

What's the fastest way to bulk insert a lot of data in SQL Server (C# client)

I am hitting some performance bottlenecks with my C# client inserting bulk data into a SQL Server 2005 database and I'm looking for ways in which to speed up the process.

I am already using the SqlClient.SqlBulkCopy (which is based on TDS) to speed up the data transfer across the wire which helped a lot, but I'm still looking for more.

I have a simple table that looks like this:

 CREATE TABLE [BulkData](  [ContainerId] [int] NOT NULL,  [BinId] [smallint] NOT NULL,  [Sequence] [smallint] NOT NULL,  [ItemId] [int] NOT NULL,  [Left] [smallint] NOT NULL,  [Top] [smallint] NOT NULL,  [Right] [smallint] NOT NULL,  [Bottom] [smallint] NOT NULL,  CONSTRAINT [PKBulkData] PRIMARY KEY CLUSTERED   (   [ContainerIdId] ASC,   [BinId] ASC,   [Sequence] ASC )) 

I'm inserting data in chunks that average about 300 rows where ContainerId and BinId are constant in each chunk and the Sequence value is 0-n and the values are pre-sorted based on the primary key.

The %Disk time performance counter spends a lot of time at 100% so it is clear that disk IO is the main issue but the speeds I'm getting are several orders of magnitude below a raw file copy.

Does it help any if I:

  1. Drop the Primary key while I am doing the inserting and recreate it later
  2. Do inserts into a temporary table with the same schema and periodically transfer them into the main table to keep the size of the table where insertions are happening small
  3. Anything else?

-- Based on the responses I have gotten, let me clarify a little bit:

Portman: I'm using a clustered index because when the data is all imported I will need to access data sequentially in that order. I don't particularly need the index to be there while importing the data. Is there any advantage to having a nonclustered PK index while doing the inserts as opposed to dropping the constraint entirely for import?

Chopeen: The data is being generated remotely on many other machines (my SQL server can only handle about 10 currently, but I would love to be able to add more). It's not practical to run the entire process on the local machine because it would then have to process 50 times as much input data to generate the output.

Jason: I am not doing any concurrent queries against the table during the import process, I will try dropping the primary key and see if that helps.

like image 979
Andrew Avatar asked Aug 23 '08 12:08

Andrew


People also ask

How can I speed up bulk insert in SQL Server?

Indexing around SQL Bulk InsertRemoving indexes prior to large inserts on a table, including when using SQL Bulk Insert, may be a best practice to increase performance.

How do I insert a lot of data into SQL?

INSERT-SELECT-UNION query to insert multiple records Thus, we can use INSERT-SELECT-UNION query to insert data into multiple rows of the table. The SQL UNION query helps to select all the data that has been enclosed by the SELECT query through the INSERT statement.

Is bulk insert faster than BCP?

BCP is faster in most cases then BULK Insert.


1 Answers

Here's how you can disable/enable indexes in SQL Server:

--Disable Index ALTER INDEX [IX_Users_UserID] SalesDB.Users DISABLE GO --Enable Index ALTER INDEX [IX_Users_UserID] SalesDB.Users REBUILD

Here are some resources to help you find a solution:

Some bulk loading speed comparisons

Use SqlBulkCopy to Quickly Load Data from your Client to SQL Server

Optimizing Bulk Copy Performance

Definitely look into NOCHECK and TABLOCK options:

Table Hints (Transact-SQL)

INSERT (Transact-SQL)

like image 194
JohnB Avatar answered Oct 02 '22 21:10

JohnB