PyLucene error with IceTea / JDK / JRE - java

I have followed the installation instructionrs http://bendemott.blogspot.de/2013/11/installing-pylucene-4-451.html for pylucene using the latest pylucene-4.9.0.0.
And when i tried to to lucene.initVM(), I get the following error:
alvas#ubi:~$ python
Python 2.7.6 (default, Mar 22 2014, 22:59:56)
[GCC 4.8.2] on linux2
Type "help", "copyright", "credits" or "license" for more information.
>>> import lucene
>>> lucene.initVM()
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGSEGV (0xb) at pc=0x00007ffba22808b8, pid=5189, tid=140718811092800
#
# JRE version: OpenJDK Runtime Environment (7.0_65-b32) (build 1.7.0_65-b32)
# Java VM: OpenJDK 64-Bit Server VM (24.65-b04 mixed mode linux-amd64 compressed oops)
# Derivative: IcedTea 2.5.3
# Distribution: Ubuntu 14.04 LTS, package 7u71-2.5.3-0ubuntu0.14.04.1
# Problematic frame:
# V [libjvm.so+0x6088b8] jni_RegisterNatives+0x58
#
# Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
# An error report file with more information is saved as:
# /home/alvas/hs_err_pid5189.log
#
# If you would like to submit a bug report, please include
# instructions on how to reproduce the bug and visit:
# http://icedtea.classpath.org/bugzilla
#
Aborted (core dumped)
And the file http://pastebin.com/6B8FyC4Z
Is there something wrong with my IceTea configuration? or my JDK? or JRE?
How should I resolve the problem?

So I took a look at your stack trace, and I don't think the issue was specifically pyLucene. In the stack trace, you see this error:
siginfo:si_signo=SIGSEGV: si_errno=0, si_code=1 (SEGV_MAPERR), si_addr=0x0000000000000000
If you look at the first part, SIGSEGV, that means you have a segmentation fault somewhere in your system. SEGV_MAPERR is the specific error, which means that OpenJDK was trying to map memory to an object and failed. This could've been caused by not enough memory, a bad pagefile/virtual memory, bad address space, or even a bad library. Why it worked on another machine could be anything. Core dumps are really useful, so if you can run
ulimit -c unlimited
that will help give you something to look at. Was this in a VM or on a physical machine? I've seen random sigsegv in my Ubuntu VMs if they don't have enough memory allocated for various Java tasks. I saw this on my ESXi hypervisors specifically, and I noticed it the most was when ESXi started to perform memory swapping. I was able to resolve this by increasing memory, rebooting the VM, and making sure my hypervisor wasn't swapping memory. Let me know if that helps. :)
Edit: I also noticed that if the underlying storage provider had poor performance, that would impact with swap data and I feel that was also am impact with sigsegv issues.

Related

How to solve the problem that mac conflicts with dl4j?

This is all error code:
19:09:34.464 [main] INFO org.nd4j.linalg.factory.Nd4jBackend - Loaded [CpuBackend] backend
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGSEGV (0xb) at pc=0x00007fff65aa2aa8, pid=2020, tid=3843
#
# JRE version: Java(TM) SE Runtime Environment (8.0_60-b27) (build 1.8.0_60-b27)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (25.60-b23 mixed mode bsd-amd64 compressed oops)
# Problematic frame:
# C [libc++abi.dylib+0x10aa8] __cxxabiv1::__si_class_type_info::has_unambiguous_public_base(__cxxabiv1::__dynamic_cast_info*, void*, int) const+0x4
#
# Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
# An error report file with more information is saved as:
# /Users/mac/Downloads/RL_DQN(19.9.27)/hs_err_pid2020.log
#
# If you would like to submit a bug report, please visit:
# http://bugreport.java.com/bugreport/crash.jsp
# The crash happened outside the Java Virtual Machine in native code.
# See problematic frame for where to report the bug.
#
How can I solve this problem?
I found that it without error in this type windows: thinkpad X1 carbon 2021; Windows 10 Enterprise Edition.
It shows error in mac:
macOS Catalina version10.15.7
MacBook Pro (13-inch, 2019, Two Thunderbolt 3 ports)
cpu 2.9 GHz doublecore Intel Core i5 memory 8GB 2133 MHz LPDDR3
graphics card Intel Iris Graphics 550 1536MB
serial number FVFZ18GLL40Y
I replace the dl4j version, but it doesn't work.
I guess mac has incompatibility with dl4j, but I don't know how to solve.
I solved this question.
This is the method: https://github.com/reactor/BlockHound/issues/37
I just replace my java version. I use this: jdk1.8.0_311
Original version is: 1.8.0_60
This was originally added to revision 4 of the question.
Make sure you are using the latest version (beta6 is pretty old at this point). Native crashes should never happen.
Please file an issue at https://gitHub.com/eclipse/deeplearning4j/issues if you still have issues after upgrading to 1.0.0-M2.

How to Use the Java Compiler Replay File to debug JVM Issues

I have a Java server process that is now intermittently crashing with the following crash report:
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGSEGV (0xb) at pc=0x00007f8c169f6df8, pid=33597, tid=140237357057792
#
# JRE version: Java(TM) SE Runtime Environment (8.0_40-b25) (build 1.8.0_40-b25)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (25.40-b25 mixed mode linux-amd64 compressed oops)
# Problematic frame:
# V [libjvm.so+0x404df8] PhaseChaitin::gather_lrg_masks(bool)+0x208
#
# Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
# An error report file with more information is saved as:
# / blah2/hs_err_pid33597.log
#
# Compiler replay data is saved as:
# /blah2/replay_pid33597.log
#
# If you would like to submit a bug report, please visit:
# http://bugreport.java.com/bugreport/crash.jsp
#
Before I think of effecting any machine configuration changes (such as changing the VM flags passed in or adding additional logging of any kind), I would like to perform as much analysis of whatever diagnostic information is available.
Does the replay_pid file have information that can help me (as an application developer) diagnose this problem or is it for VM crash reporting to Oracle? Are there any tools available that can analyze it (similar to a thread dump analyzer for instance)?
Any help/hints appreciated.
Thanks
AD
Yes, the replay file can help you to reproduce the issue in a controlled manner.
There's an excellent presentation "Analyzing HotSpot Crashes" by Volker Simonis which covers this - here's a direct link to the relevant example in that talk: https://www.youtube.com/watch?v=xC8fEeo7izI&feature=youtu.be&t=1888

Why a JRE crashes because of core dumping?

My JRE crashed when executing a shared library function called from java code via JNI. The output says that JRE crashed because it "Failed to write core dump. Core dumps have been disabled.".
After many days googling without explanation, my questions is:
Where can I find info about JRE core dumping? I want to understand the problem and the solution.
I know the output recommends executing "ulimit -c unlimited" and this questions proposes solutions (How to enable program to dump core on linux?)
Here I've pasted the JRE output:
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGSEGV (0xb) at pc=0x00007f7e541cec55, pid=20390, tid=140180586714880
#
# JRE version: Java(TM) SE Runtime Environment (7.0_55-b13) (build 1.7.0_55-b13)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (24.55-b03 mixed mode linux-amd64 compressed oops)
# Problematic frame:
# C [libXXX.so+0x1fc55] NewMask_UnsetAll+0x15
#
# Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
# An error report file with more information is saved as:
# /XXX/hs_err_pid20390.log
#
# If you would like to submit a bug report, please visit:
# http://bugreport.sun.com/bugreport/crash.jsp
#
I know this is an old question…
This message "Failed to write core dump. Core dumps have been disabled." didn't mean this was the cause of the crash. The word "Failed" is indeed misleading here.
This message only means that core dumps are disabled in the Java VM (which is usually the case), so that such dumps are not produced in case of crash.

Java Runtime Environment SIGSEGV error on server startup

I'm using java version "1.7.0_45" with eclipse kepler and on server startup I'm getting the below error log.
Although I've found several posts[1, 2] regarding the same issue, I've tried everything from adding -Dorg.eclipse.swt.browser.DefaultType=mozilla, -XX:LoopUnrollLimit=1 and ulimit -c unlimited but nothing worked for me.
Is there any other work around ?
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGSEGV (0xb) at pc=0x0000000000000000, pid=7084, tid=139749936641792
#
# JRE version: Java(TM) SE Runtime Environment (7.0_45-b18) (build 1.7.0_45-b18)
# Java VM: Java HotSpot(TM) 64-Bit Server VM (24.45-b08 mixed mode linux-amd64 compressed oops)
# Problematic frame:
# C 0x0000000000000000
#
# Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
#
SIGSEV means that your program is producing a so called segmentation fault. Writing over array boundaries, OR invalid memory addresses, etc... So I think you have to re install eclipse first and then see if you have to reinstall java and eclipse both... if the first wont work for you.
I am not sure if this is the correct guess, I hope no one downvotes this...
I had a similar error while I started netbeans in Ubuntu 13.04. I fixed it with
sudo apt-get install openjdk-7-jdk
Try and see if it works.

Hadoop dies on ld-linux.so

I have the following setup:
Hadoop 1.2.1
Oracle Java 1.7
Suse Enterprise Server 10 32bit
If I execute the Pi-example in standalone mode with
bin/hadoop jar hadoop-examples-1.2.1.jar pi 10 10
then Java dies the hard way, telling me
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGFPE (0x8) at pc=0xb7efa20b, pid=9494, tid=3070639008
#
# JRE version: Java(TM) SE Runtime Environment (7.0_40-b43) (build 1.7.0_40-b43)
# Java VM: Java HotSpot(TM) Server VM (24.0-b56 mixed mode linux-x86 )
# Problematic frame:
# C [ld-linux.so.2+0x920b] do_lookup_x+0xab
#
# Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
# An error report file with more information is saved as:
# /opt/hadoop-1.2.1-new/hs_err_pid9494.log
#
# If you would like to submit a bug report, please visit:
# http://bugreport.sun.com/bugreport/crash.jsp
# The crash happened outside the Java Virtual Machine in native code.
# See problematic frame for where to report the bug.
#
(The full trace is here)
On a distributed setup, i can start-all components and they are ideling fine. But when I submit a job, then the jobtracker dies immediately with an java.io.EOFException, I assume this is due to the same error as above.
I have already tried the same hadoop on another computer, there everything is fine (altough this one runs Arch Linux 64bit), and other Javas (OpenJDK, 1.6, 1.7) don't help.
Any suggestions?
Probably Hadoop includes a native library that was either compiled for a different platform (e.g. 64 bit instead of 32 bit), or the library expects a different environment. The stack trace also shows that JVM_LoadLibrary() is trying to load a native lib.
Make sure you downloaded the correct version of Hadoop for your platform, or compile it yourself for your target platform.

Categories

Resources