Back to Multiple platform build/check report for BioC 3.19:   simplified   long
ABCDEFGHIJKLMNOPQ[R]STUVWXYZ

This page was generated on 2024-05-09 11:41:21 -0400 (Thu, 09 May 2024).

HostnameOSArch (*)R versionInstalled pkgs
nebbiolo1Linux (Ubuntu 22.04.3 LTS)x86_644.4.0 (2024-04-24) -- "Puppy Cup" 4748
palomino3Windows Server 2022 Datacenterx644.4.0 (2024-04-24 ucrt) -- "Puppy Cup" 4484
lconwaymacOS 12.7.1 Montereyx86_644.4.0 (2024-04-24) -- "Puppy Cup" 4514
kunpeng2Linux (openEuler 22.03 LTS-SP1)aarch644.4.0 beta (2024-04-15 r86425) -- "Puppy Cup" 4480
Click on any hostname to see more info about the system (e.g. compilers)      (*) as reported by 'uname -p', except on Windows and Mac OS X

Package 1752/2300HostnameOS / ArchINSTALLBUILDCHECKBUILD BIN
RGMQL 1.24.0  (landing page)
Simone Pallotta
Snapshot Date: 2024-05-08 14:00:19 -0400 (Wed, 08 May 2024)
git_url: https://git.bioconductor.org/packages/RGMQL
git_branch: RELEASE_3_19
git_last_commit: 953c821
git_last_commit_date: 2024-04-30 11:03:12 -0400 (Tue, 30 Apr 2024)
nebbiolo1Linux (Ubuntu 22.04.3 LTS) / x86_64  OK    OK    WARNINGS  UNNEEDED, same version is already published
palomino3Windows Server 2022 Datacenter / x64  OK    OK    ERROR    OK  
lconwaymacOS 12.7.1 Monterey / x86_64... NOT SUPPORTED ...
kunpeng2Linux (openEuler 22.03 LTS-SP1) / aarch64  OK    ERROR  skipped
kjohnson3macOS 13.6.5 Ventura / arm64see weekly results here

BUILD results for RGMQL on kunpeng2


To the developers/maintainers of the RGMQL package:
- Allow up to 24 hours (and sometimes 48 hours) for your latest push to git@git.bioconductor.org:packages/RGMQL.git to reflect on this report. See Troubleshooting Build Report for more information.
- Use the following Renviron settings to reproduce errors and warnings.
- If 'R CMD check' started to fail recently on the Linux builder(s) over a missing dependency, add the missing dependency to 'Suggests:' in your DESCRIPTION file. See Renviron.bioc for more information.
- See Martin Grigorov's blog post for how to debug Linux ARM64 related issues on a x86_64 host.

raw results


Summary

Package: RGMQL
Version: 1.24.0
Command: /home/biocbuild/R/R-beta-2024-04-15_r86425/bin/R CMD build --keep-empty-dirs --no-resave-data RGMQL
StartedAt: 2024-05-09 02:58:35 -0000 (Thu, 09 May 2024)
EndedAt: 2024-05-09 02:59:39 -0000 (Thu, 09 May 2024)
EllapsedTime: 64.1 seconds
RetCode: 1
Status:   ERROR  
PackageFile: None
PackageFileSize: NA

Command output

##############################################################################
##############################################################################
###
### Running command:
###
###   /home/biocbuild/R/R-beta-2024-04-15_r86425/bin/R CMD build --keep-empty-dirs --no-resave-data RGMQL
###
##############################################################################
##############################################################################


* checking for file ‘RGMQL/DESCRIPTION’ ... OK
* preparing ‘RGMQL’:
* checking DESCRIPTION meta-information ... OK
* installing the package to build vignettes
* creating vignettes ... ERROR
--- re-building ‘RGMQL-vignette.Rmd’ using rmarkdown
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
24/05/09 02:59:37 INFO SparkContext: Running Spark version 2.2.0
24/05/09 02:59:37 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
24/05/09 02:59:38 INFO SparkContext: Submitted application: GMQL-R
24/05/09 02:59:38 INFO SecurityManager: Changing view acls to: biocbuild
24/05/09 02:59:38 INFO SecurityManager: Changing modify acls to: biocbuild
24/05/09 02:59:38 INFO SecurityManager: Changing view acls groups to: 
24/05/09 02:59:38 INFO SecurityManager: Changing modify acls groups to: 
24/05/09 02:59:38 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(biocbuild); groups with view permissions: Set(); users  with modify permissions: Set(biocbuild); groups with modify permissions: Set()
24/05/09 02:59:38 INFO PlatformDependent: Your platform does not provide complete low-level API for accessing direct buffers reliably. Unless explicitly requested, heap buffer will always be preferred to avoid potential system unstability.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 WARN Utils: Service 'sparkDriver' could not bind on a random free port. You may check whether configuring an appropriate binding address.
24/05/09 02:59:38 ERROR SparkContext: Error initializing SparkContext.
java.net.BindException: Cannot assign requested address: Service 'sparkDriver' failed after 16 retries (on a random free port)! Consider explicitly setting the appropriate binding address for the service 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the correct binding address.
	at java.base/sun.nio.ch.Net.bind0(Native Method)
	at java.base/sun.nio.ch.Net.bind(Net.java:555)
	at java.base/sun.nio.ch.ServerSocketChannelImpl.netBind(ServerSocketChannelImpl.java:337)
	at java.base/sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:294)
	at io.netty.channel.socket.nio.NioServerSocketChannel.doBind(NioServerSocketChannel.java:127)
	at io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:501)
	at io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1218)
	at io.netty.channel.AbstractChannelHandlerContext.invokeBind(AbstractChannelHandlerContext.java:496)
	at io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:481)
	at io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:965)
	at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:210)
	at io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:353)
	at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:399)
	at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:446)
	at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:131)
	at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:144)
	at java.base/java.lang.Thread.run(Thread.java:833)
24/05/09 02:59:38 INFO SparkContext: Successfully stopped SparkContext

Quitting from lines  at lines 250-251 [init] (RGMQL-vignette.Rmd)
Error: processing vignette 'RGMQL-vignette.Rmd' failed with diagnostics:
java.net.BindException: Cannot assign requested address: Service 'sparkDriver' failed after 16 retries (on a random free port)! Consider explicitly setting the appropriate binding address for the service 'sparkDriver' (for example spark.driver.bindAddress for SparkDriver) to the correct binding address.
--- failed re-building ‘RGMQL-vignette.Rmd’

SUMMARY: processing the following file failed:
  ‘RGMQL-vignette.Rmd’

Error: Vignette re-building failed.
Execution halted