[OpenStack-I18n] [OpenStack-Infra] translate-dev wildfly/zanata/trove issues
Ian Y. Choi
ianyrchoi at gmail.com
Mon Jan 16 04:15:31 UTC 2017
Thanks a lot for such sharing Ian! - not me :)
I am copying to i18n mailing list and Zanata dev team for FYI purpose
(as what Kato-san kindly RST markups for option on another mail thread).
Note that this issue is re-generatable: I am able recreate the issue on
translate-dev
: When I create a new version from openstack-manuals from master branch
[1] - 20107950 words,
there is no further web responses from translate-dev.o.o around after
3/4 of total words were processed
like [2-4]. It was completely fine for a version creation with 2610150
words.
If someone has a good idea to further investigate on it, please share to
us :)
[1]
https://translate-dev.openstack.org/iteration/view/openstack-manuals/master
[2]
https://translate-dev.openstack.org/iteration/view/openstack-manuals/stable-newton
[3]
https://translate-dev.openstack.org/iteration/view/openstack-manuals/test-version-creation
[4]
https://translate-dev.openstack.org/iteration/view/openstack-manuals/test-version-creation2
With many thanks,
/Ian
Ian Wienand wrote on 1/16/2017 12:03 PM:
> Hi,
>
> I was alerted to translate-dev performance issues today. Indeed, it
> seemed that things were going crazy with the java wildfly process
> sucking up all CPU.
>
> At first there didn't seem to be anything in the logs. Java was
> clearly going mad however, with the following threads going flat-out.
>
> --- gone crazy processes ---
> 14807 wildfly 20 0 9.892g 4.729g 34232 R 93.7 60.7 26:56.73 java
> 14804 wildfly 20 0 9.892g 4.729g 34232 R 92.4 60.7 26:51.91 java
> 14806 wildfly 20 0 9.892g 4.729g 34232 R 92.4 60.7 26:53.92 java
> 14808 wildfly 20 0 9.892g 4.729g 34232 R 92.4 60.7 26:53.97 java
> 14810 wildfly 20 0 9.892g 4.729g 34232 R 92.4 60.7 26:56.28 java
> 14809 wildfly 20 0 9.892g 4.729g 34232 R 92.1 60.7 26:57.74 java
> 14803 wildfly 20 0 9.892g 4.729g 34232 R 91.1 60.7 26:54.90 java
> 14805 wildfly 20 0 9.892g 4.729g 34232 R 90.4 60.7 26:52.44 java
> ---
>
> Hoping to find a easy smoking-gun, I made the java process dump it's
> threads to see what these are doing
>
> --- thread dump ---
>
> 14807 14804 14806 14808 14810 14809 14803, 14805
> 0x39d7 0x39d4 0x39d6 0x39d8 0x39da 0x39d9 0x39d3 0x39d5
>
> "GC task thread#4 (ParallelGC)" os_prio=0 tid=0x00007fb2a8026000 nid=0x39d7 runnable
> "GC task thread#4 (ParallelGC)" os_prio=0 tid=0x00007fb2a8026000 nid=0x39d7 runnable
> "GC task thread#4 (ParallelGC)" os_prio=0 tid=0x00007fb2a8026000 nid=0x39d7 runnable
> "GC task thread#4 (ParallelGC)" os_prio=0 tid=0x00007fb2a8026000 nid=0x39d7 runnable
>
> "GC task thread#1 (ParallelGC)" os_prio=0 tid=0x00007fb2a8021000 nid=0x39d4 runnable
> "GC task thread#1 (ParallelGC)" os_prio=0 tid=0x00007fb2a8021000 nid=0x39d4 runnable
> "GC task thread#1 (ParallelGC)" os_prio=0 tid=0x00007fb2a8021000 nid=0x39d4 runnable
> "GC task thread#1 (ParallelGC)" os_prio=0 tid=0x00007fb2a8021000 nid=0x39d4 runnable
>
> "GC task thread#3 (ParallelGC)" os_prio=0 tid=0x00007fb2a8024800 nid=0x39d6 runnable
> "GC task thread#3 (ParallelGC)" os_prio=0 tid=0x00007fb2a8024800 nid=0x39d6 runnable
> "GC task thread#3 (ParallelGC)" os_prio=0 tid=0x00007fb2a8024800 nid=0x39d6 runnable
> "GC task thread#3 (ParallelGC)" os_prio=0 tid=0x00007fb2a8024800 nid=0x39d6 runnable
>
> "GC task thread#5 (ParallelGC)" os_prio=0 tid=0x00007fb2a8028000 nid=0x39d8 runnable
> "GC task thread#5 (ParallelGC)" os_prio=0 tid=0x00007fb2a8028000 nid=0x39d8 runnable
> "GC task thread#5 (ParallelGC)" os_prio=0 tid=0x00007fb2a8028000 nid=0x39d8 runnable
> "GC task thread#5 (ParallelGC)" os_prio=0 tid=0x00007fb2a8028000 nid=0x39d8 runnable
>
> "GC task thread#7 (ParallelGC)" os_prio=0 tid=0x00007fb2a802b800 nid=0x39da runnable
> "GC task thread#7 (ParallelGC)" os_prio=0 tid=0x00007fb2a802b800 nid=0x39da runnable
> "GC task thread#7 (ParallelGC)" os_prio=0 tid=0x00007fb2a802b800 nid=0x39da runnable
> "GC task thread#7 (ParallelGC)" os_prio=0 tid=0x00007fb2a802b800 nid=0x39da runnable
>
> "GC task thread#6 (ParallelGC)" os_prio=0 tid=0x00007fb2a8029800 nid=0x39d9 runnable
> "GC task thread#6 (ParallelGC)" os_prio=0 tid=0x00007fb2a8029800 nid=0x39d9 runnable
> "GC task thread#6 (ParallelGC)" os_prio=0 tid=0x00007fb2a8029800 nid=0x39d9 runnable
> "GC task thread#6 (ParallelGC)" os_prio=0 tid=0x00007fb2a8029800 nid=0x39d9 runnable
>
> "GC task thread#0 (ParallelGC)" os_prio=0 tid=0x00007fb2a801f000 nid=0x39d3 runnable
> "GC task thread#0 (ParallelGC)" os_prio=0 tid=0x00007fb2a801f000 nid=0x39d3 runnable
> "GC task thread#0 (ParallelGC)" os_prio=0 tid=0x00007fb2a801f000 nid=0x39d3 runnable
> "GC task thread#0 (ParallelGC)" os_prio=0 tid=0x00007fb2a801f000 nid=0x39d3 runnable
> ---
>
> Unfortunately, these all appear to be GC threads so there's nothing
> really obvious there.
>
> However, eventually in the log you start getting stuff like
>
> ---
> 2017-01-16T01:58:00,844Z SEVERE [javax.enterprise.resource.webcontainer.jsf.application] (default task-43) Error Rendering View[/project/project.xhtml]: javax.el.ELException: /WEB-INF/layout/project/settings-tab-languages.xhtml @117,88 rendered="#{not projectHome.hasLocaleAlias(locale)}": javax.persistence.PersistenceException: org.hibernate.HibernateException: Transaction was rolled back in a different thread!
> ...
> 2017-01-16T01:08:48,805Z WARN [org.hibernate.engine.jdbc.spi.SqlExceptionHelper] (pool-5-thread-1) SQL Error: 0, SQLState: null
> 2017-01-16T01:08:48,806Z ERROR [org.hibernate.engine.jdbc.spi.SqlExceptionHelper] (pool-5-thread-1) javax.resource.ResourceException: IJ000460: Error checking for a transaction
> 2017-01-16T01:08:48,813Z ERROR [org.hibernate.AssertionFailure] (pool-5-thread-1) HHH000099: an assertion failure occured (this may indicate a bug in Hibernate, but is more likely due to unsafe use of the session): org.hibernate.exception.GenericJDBCException: Could not open connection
> 2017-01-16T01:08:48,813Z WARN [com.arjuna.ats.jta] (pool-5-thread-1) ARJUNA016029: SynchronizationImple.afterCompletion - failed for org.hibernate.engine.transaction.synchronization.internal.RegisteredSynchronization at 3fc123aa with exception: org.hibernate.AssertionFailure: Exception releasing cache locks
> ---
>
> There's literally hundreds of lines of traceback across various
> exceptions, but at the bottom this really seems to point to mysql
> being at the root of all this?
>
> In this case, mysql is running in RAX's cloud db service. I logged
> into the RAX control-panel and everything for the DB was green. I
> used the same credentials as zanata to authenticate and I can connect
> to the db via translate-dev.o.o and do a few basic things.
>
> ---
> MySQL [(none)]> USE zanata;
> Reading table information for completion of table and column names
> You can turn off this feature to get a quicker startup with -A
>
> Database changed
> MySQL [zanata]> show tables;
> +----------------------------------+
> | Tables_in_zanata |
> +----------------------------------+
> | Activity |
> | DATABASECHANGELOG |
> ...
> ---
>
> So it's not as simple as "db not working" :( I have copied the console
> log to [1] in case that makes any sense to zanata people.
>
> If this persists, and isn't a known Zanata issue, I think we might
> have to enable root access for the db via the API which might allow us
> to do some thing like debug long-queries. Otherwise, there might be
> something in the mysql logs available to RAX support people that would
> otherwise help identify the issue. OTOH, maybe it's as simple as
> upping java heaps or something.
>
> In the mean time, I have restarted wildfly which seems to get things
> going for a while.
>
> -i
>
> [1] http://people.redhat.com/~iwienand/console.log.translate-dev-2017-01-16.log.gz
>
> _______________________________________________
> OpenStack-Infra mailing list
> OpenStack-Infra at lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-infra
More information about the OpenStack-I18n
mailing list