NetworkCrawler.java
- /* Copyright 2002-2013 CS Systèmes d'Information
- * Licensed to CS Systèmes d'Information (CS) under one or more
- * contributor license agreements. See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * CS licenses this file to You under the Apache License, Version 2.0
- * (the "License"); you may not use this file except in compliance with
- * the License. You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
- package org.orekit.data;
- import java.io.File;
- import java.io.IOException;
- import java.io.InputStream;
- import java.net.URISyntaxException;
- import java.net.URL;
- import java.net.URLConnection;
- import java.text.ParseException;
- import java.util.ArrayList;
- import java.util.List;
- import java.util.regex.Matcher;
- import java.util.regex.Pattern;
- import java.util.zip.GZIPInputStream;
- import org.apache.commons.math3.exception.util.DummyLocalizable;
- import org.orekit.errors.OrekitException;
- /** Provider for data files directly fetched from network.
- * <p>
- * This class handles a list of URLs pointing to data files or zip/jar on
- * the net. Since the net is not a tree structure the list elements
- * cannot be top elements recursively browsed as in {@link
- * DirectoryCrawler}, they must be data files or zip/jar archives.
- * </p>
- * <p>
- * The files fetched from network can be locally cached on disk. This prevents
- * too frequent network access if the URLs are remote ones (for example
- * original internet URLs).
- * </p>
- * <p>
- * If the URL points to a remote server (typically on the web) on the other side
- * of a proxy server, you need to configure the networking layer of your
- * application to use the proxy. For a typical authenticating proxy as used in
- * many corporate environments, this can be done as follows using for example
- * the AuthenticatorDialog graphical authenticator class that can be found
- * in the tests directories:
- * <pre>
- * System.setProperty("http.proxyHost", "proxy.your.domain.com");
- * System.setProperty("http.proxyPort", "8080");
- * System.setProperty("http.nonProxyHosts", "localhost|*.your.domain.com");
- * Authenticator.setDefault(new AuthenticatorDialog());
- * </pre>
- * </p>
- * <p>
- * Gzip-compressed files are supported.
- * </p>
- * <p>
- * Zip archives entries are supported recursively.
- * </p>
- * <p>
- * This is a simple application of the <code>visitor</code> design pattern for
- * list browsing.
- * </p>
- * @see DataProvidersManager
- * @author Luc Maisonobe
- */
- public class NetworkCrawler implements DataProvider {
- /** URLs list. */
- private final List<URL> urls;
- /** Connection timeout (milliseconds). */
- private int timeout;
- /** Build a data classpath crawler.
- * <p>The default timeout is set to 10 seconds.</p>
- * @param urls list of data file URLs
- */
- public NetworkCrawler(final URL... urls) {
- this.urls = new ArrayList<URL>();
- for (final URL url : urls) {
- this.urls.add(url);
- }
- timeout = 10000;
- }
- /** Set the timeout for connection.
- * @param timeout connection timeout in milliseconds
- */
- public void setTimeout(final int timeout) {
- this.timeout = timeout;
- }
- /** {@inheritDoc} */
- public boolean feed(final Pattern supported, final DataLoader visitor)
- throws OrekitException {
- try {
- OrekitException delayedException = null;
- boolean loaded = false;
- for (URL url : urls) {
- try {
- if (visitor.stillAcceptsData()) {
- final String name = url.toURI().toString();
- final String fileName = new File(url.getPath()).getName();
- if (ZIP_ARCHIVE_PATTERN.matcher(fileName).matches()) {
- // browse inside the zip/jar file
- new ZipJarCrawler(url).feed(supported, visitor);
- loaded = true;
- } else {
- // remove suffix from gzip files
- final Matcher gzipMatcher = GZIP_FILE_PATTERN.matcher(fileName);
- final String baseName = gzipMatcher.matches() ? gzipMatcher.group(1) : fileName;
- if (supported.matcher(baseName).matches()) {
- final InputStream stream = getStream(url);
- // visit the current file
- if (gzipMatcher.matches()) {
- visitor.loadData(new GZIPInputStream(stream), name);
- } else {
- visitor.loadData(stream, name);
- }
- stream.close();
- loaded = true;
- }
- }
- }
- } catch (OrekitException oe) {
- // maybe the next path component will be able to provide data
- // wait until all components have been tried
- delayedException = oe;
- }
- }
- if (!loaded && delayedException != null) {
- throw delayedException;
- }
- return loaded;
- } catch (URISyntaxException use) {
- throw new OrekitException(use, new DummyLocalizable(use.getMessage()));
- } catch (IOException ioe) {
- throw new OrekitException(ioe, new DummyLocalizable(ioe.getMessage()));
- } catch (ParseException pe) {
- throw new OrekitException(pe, new DummyLocalizable(pe.getMessage()));
- }
- }
- /** Get the stream to read from the remote URL.
- * @param url url to read from
- * @return stream to read the content of the URL
- * @throws IOException if the URL cannot be opened for reading
- */
- private InputStream getStream(final URL url) throws IOException {
- final URLConnection connection = url.openConnection();
- connection.setConnectTimeout(timeout);
- return connection.getInputStream();
- }
- }