Hi Sean, It's true that the sbt is trying different links but ALL of them have connections issue (which is actually 404 File not found error) and the build process takes forever connecting different links..
I don't think it's a proxy issue because my other projects (other than spark) builds well and I can wget things like wget www.yahoo.com. I am surprised that there is not much noise raised on this issue --- at least on the most recent spark release. Best, Leo On Wed, Apr 16, 2014 at 12:46 PM, Sean Owen <so...@cloudera.com> wrote: > This is just a red herring. You are seeing the build fail to contact > many repos it knows about, including ones that do not have a given > artifact. > > This is almost always a symptom of network connectivity problem, like > perhaps a proxy in between, esp. one that breaks HTTPS connections. > You may need to fix your proxy settings for Maven locally. > -- > Sean Owen | Director, Data Science | London > > > On Wed, Apr 16, 2014 at 8:32 PM, Arpit Tak <arpi...@sigmoidanalytics.com> > wrote: > > Its because , there is no sl4f directory exists there.... may be they > > updating it ..... > > https://oss.sonatype.org/content/repositories/snapshots/org/ > > > > Hard luck .... try after some time... > > > > Regards, > > Arpit > > > > > > On Thu, Apr 17, 2014 at 12:33 AM, Yiou Li <liy...@gmail.com> wrote: > >> > >> Hi all, > >> > >> I am trying to build spark assembly using sbt and got connection error > >> when resolving dependencies: > >> > >> I tried web browser and wget some of the dependency links in the error > and > >> also got 404 error too. > >> > >> This happened to the following branches: > >> spark-0.8.1-incubating > >> spark-0.9.1 > >> spark-0.9.1-bin-hadoop2 > >> > >> Can somebody please kindly advise? > >> > >> Best, > >> Leo > >> > >> > >> Launching sbt from sbt/sbt-launch-0.12.4.jar > >> [info] Loading project definition from > >> > /home/leo/workspace_spark/spark-core/spark-0.9.1-bin-hadoop2/project/project > >> [info] Loading project definition from > >> /home/leo/workspace_spark/spark-core/spark-0.9.1-bin-hadoop2/project > >> [info] Set current project to root (in build > >> file:/home/leo/workspace_spark/spark-core/spark-0.9.1-bin-hadoop2/) > >> [info] Updating > >> > {file:/home/leo/workspace_spark/spark-core/spark-0.9.1-bin-hadoop2/}core... > >> [info] Resolving org.slf4j#slf4j-log4j12;1.7.2 ... > >> [error] Server access Error: Connection timed out > >> url= > https://oss.sonatype.org/content/repositories/snapshots/org/slf4j/slf4j-log4j12/1.7.2/slf4j-log4j12-1.7.2.pom > >> [error] Server access Error: Connection timed out > >> url= > https://oss.sonatype.org/service/local/staging/deploy/maven2/org/slf4j/slf4j-log4j12/1.7.2/slf4j-log4j12-1.7.2.pom > >> [info] Resolving commons-daemon#commons-daemon;1.0.10 ... > >> [error] Server access Error: Connection timed out > >> url= > https://oss.sonatype.org/content/repositories/snapshots/commons-daemon/commons-daemon/1.0.10/commons-daemon-1.0.10.pom > >> [error] Server access Error: Connection timed out > >> url= > https://oss.sonatype.org/service/local/staging/deploy/maven2/commons-daemon/commons-daemon/1.0.10/commons-daemon-1.0.10.pom > >> [error] Server access Error: Connection timed out > >> url= > https://repository.cloudera.com/artifactory/cloudera-repos/commons-daemon/commons-daemon/1.0.10/commons-daemon-1.0.10.pom > >> [info] Resolving org.apache.commons#commons-parent;23 ... > >> [error] Server access Error: Connection timed out > >> url= > https://oss.sonatype.org/content/repositories/snapshots/org/apache/commons/commons-parent/23/commons-parent-23.pom > >> > >> (truncated) > > > > >