Redian新闻
>
这VM CPU太weak怎么Identify?
avatar
这VM CPU太weak怎么Identify?# Java - 爪哇娇娃
l*u
1
I am going to import very large stock market data from multiple .txt files (
over 3000 files) into sql database. I was wondering which is the most
efficient way to do that?
I used bulk insert and it takes me forever time (estimated 7 days)
avatar
r*s
2
Perf问题,
我总觉着是VM CPU太weak,
可能Ops放太多VM share一个CPU,
怎么Identify这个问题?
至少让我不用担心VM的问题?
现在那CPU动不动就99.9%.
avatar
B*g
3
SSIS

(

【在 l****u 的大作中提到】
: I am going to import very large stock market data from multiple .txt files (
: over 3000 files) into sql database. I was wondering which is the most
: efficient way to do that?
: I used bulk insert and it takes me forever time (estimated 7 days)

avatar
g*g
4
动不动就99.9%不就证明CPU不够给力了吗?一般cpu monitoring是通过脚本,用另外的
机器来管理的。
常用的命令是top。

【在 r*****s 的大作中提到】
: Perf问题,
: 我总觉着是VM CPU太weak,
: 可能Ops放太多VM share一个CPU,
: 怎么Identify这个问题?
: 至少让我不用担心VM的问题?
: 现在那CPU动不动就99.9%.

avatar
j*n
5
No dude, SSIS is slower than bcp...
if I am doing this, i'd like to have dba to extend the db size to at least
100GB more with multiple data files on different disks; also i'd like to set
up db to simple recovery model, and set bcp commitment to every 500,000
rows; oh remember remove all indexes from the table first.
anything else?

【在 B*****g 的大作中提到】
: SSIS
:
: (

avatar
e*t
6
没啥好办法。
让ops尝试把其他的VM换到别的node去,如果还是不能改善你application的
performance,那就只有2条路:
1. scale vertically, 上更强劲的新机器.
2. scale horizontally, 上更多的机器,增加你application的throughput.

【在 r*****s 的大作中提到】
: Perf问题,
: 我总觉着是VM CPU太weak,
: 可能Ops放太多VM share一个CPU,
: 怎么Identify这个问题?
: 至少让我不用担心VM的问题?
: 现在那CPU动不动就99.9%.

avatar
p*l
7
Pretty much these things. Specifying batchsize is important.

set

【在 j*****n 的大作中提到】
: No dude, SSIS is slower than bcp...
: if I am doing this, i'd like to have dba to extend the db size to at least
: 100GB more with multiple data files on different disks; also i'd like to set
: up db to simple recovery model, and set bcp commitment to every 500,000
: rows; oh remember remove all indexes from the table first.
: anything else?

avatar
e*t
8
他这个就在hypervisor上,直接用hypervisor的UI看就好了。

【在 g*****g 的大作中提到】
: 动不动就99.9%不就证明CPU不够给力了吗?一般cpu monitoring是通过脚本,用另外的
: 机器来管理的。
: 常用的命令是top。

avatar
c*t
9
1. Bulk Insert is faster than SSIS if you turn on the table lock.
2. If you can make the data in all files are sorted, you can save the time
on index rebuild.
3. No matter Simple Recovery or Bulk Mode, SQL Server will start writing
huge data to log file after the first file, the size usually be 2~3 times
the data you are importing, so I suggest you shrink the log file after each
bulk-insert each file. (MSSQL will not write too much data to the log file
if the table is empty)
avatar
i*c
10
write a class,do sql insert only
avatar
j*7
11
I think the most important is to load the files in parallel, other
considerations include:
set DB recovery mode to be Simple
pre-allocate database files and optimize the disk system
are you going to load to a single table? can consider a partitioned design
perhaps.
avatar
G*d
12
maybe because of tons of reason
RAID?
recovery model?
even with SSIS, it should not take such long time...
check your log/locks and if you have profiler running you can monitor it.
avatar
c*o
13
SAS is a way to go

(

【在 l****u 的大作中提到】
: I am going to import very large stock market data from multiple .txt files (
: over 3000 files) into sql database. I was wondering which is the most
: efficient way to do that?
: I used bulk insert and it takes me forever time (estimated 7 days)

avatar
c*y
14
删除index,用BCP
avatar
W*9
15
DTS, easy and fast
avatar
w*e
16
Now, it is SSIS....

【在 W******9 的大作中提到】
: DTS, easy and fast
相关阅读
logo
联系我们隐私协议©2024 redian.news
Redian新闻
Redian.news刊载任何文章,不代表同意其说法或描述,仅为提供更多信息,也不构成任何建议。文章信息的合法性及真实性由其作者负责,与Redian.news及其运营公司无关。欢迎投稿,如发现稿件侵权,或作者不愿在本网发表文章,请版权拥有者通知本网处理。