-
Notifications
You must be signed in to change notification settings - Fork 18
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Segmentation Fault when running the sample code #2
Comments
Hi, could you present the hosts.xml information and your compilation options to me (cmake -Dxxxx=xx) ? |
The hosts.xml is like this:
compilation options are the same as in your README: The following is the the result when I ran
I run the python script on nerv1, which is the first machine in hosts.xml. Thanks! |
Hi, thanks for your information. |
Hi Thanks for the reply! This segfaults issue is solved. But I still cannot run the code well. I am trying to debug. This is my current output when run the script: [chao@nerv1 scripts]$ ./run2.py config.xml noccocc "-t 24 -c 10 -r 100" bank 3 [START] Input parsing done. This "Connect Memory Region failed error" is due to ibv_reg_mr() returns null. Do you have any idea why ibv_reg_mr() function can fail? I appreciate your advise. |
Hi, it seems that there is no 2M huge pages available on your machine (This results in more memory to register the memory region on the NIC). ps: If you can not use huge page, you can configure the RNIC according to this post https://community.mellanox.com/docs/DOC-1120 to allow RNIC register larger memory. But huge page is suggested for better performance. |
I allocated enough huge page memory as your suggested. And the bank bench is now working fine on three machines and with 8 threads on each machine. Thanks for all the help :) |
Hi,
I am trying to run the sample script but I encounter a Segmentation Fault error. Do you have any suggestions on resolving this issue?
I used the hosts.xml file and config.xml file as mentioned in the README. Here is the command I use:
./run2.py config.xml noccocc "-t 24 -c 10 -r 100" tpcc 3
The output is like this: (I added some more output on my own):
[START] Input parsing done.
[START] cleaning remaining processes.
ssh -n -f nerv2 "cd /home/chao/git_repos/rocc/scripts/ && rm log"
ssh -n -f nerv2 "cd /home/chao/git_repos/rocc/scripts/ && ./noccocc --bench tpcc --txn-flags 1 --verbose --config config.
xml --id 1 -t 24 -c 10 -r 100 -p 3 1>log 2>&1 &"
ssh -n -f nerv3 "cd /home/chao/git_repos/rocc/scripts/ && rm log"
ssh -n -f nerv3 "cd /home/chao/git_repos/rocc/scripts/ && ./noccocc --bench tpcc --txn-flags 1 --verbose --config config.
xml --id 2 -t 24 -c 10 -r 100 -p 3 1>log 2>&1 &"
cd /home/chao/git_repos/rocc/scripts/ && ./noccocc --bench tpcc --txn-flags 1 --verbose --config config.xml --id 0 -t 24
-c 10 -r 100 -p 3
NOCC started with program [noccocc]. at 06-09-2018 09:36:48
[tpcc] settings:
new_order_remote_item_pct : 1
uniform_item_dist : 0
micro dist :20
[bench_runner.cc:324] Use TCP port 8888
[bench_runner.cc:346] use scale factor: 72; with total 24 threads.
[view.h:48] Start with 0 backups.
[view.cc:10] total 3 backups to assign
Txn NewOrder, 100
Remote counts: 100
NAIVE: 4[util.cc:164] huge page alloc failed!
[librdma] get device name mlx4_0, idx 0
[librdma] : Device 0 has 1 ports
[bench_runner.cc:153] Total logger area 0.00585938G.
[bench_runner.cc:163] add RDMA store size 4.88281G.
[bench_runner.cc:172] [Mem] RDMA heap size 8.03902G.
[util.cc:164] huge page alloc failed!
[util.cc:164] huge page alloc failed!
[NOCC] Meet a segmentation fault!
stack trace:
./noccocc() [0x4b3bb8]
/lib64/libc.so.6 : ()+0x35270
/lib64/libc.so.6 : ()+0x8981d
./noccocc : MemDB::AddSchema(int, TABLE_CLASS, int, int, int, int, bool)+0x105
./noccocc : nocc::oltp::tpcc::TpccMainRunner::init_store(MemDB&)+0xe0
./noccocc : nocc::oltp::BenchRunner::run()+0x3d4
./noccocc : nocc::oltp::tpcc::TpccTest(int, char*)+0x143
./noccocc : main()+0x589
/lib64/libc.so.6 : __libc_start_main()+0xf5
./noccocc() [0x47813c]
Thanks!
The text was updated successfully, but these errors were encountered: