jose4j / JWT Examples

本文介绍如何使用JOSE4J库生成及验证JSON Web Token (JWT),包括使用RSA密钥对JWT进行签名和验证的过程,通过HTTPS JWK端点获取公钥,使用JWKs进行验证,以及X.509证书的处理方式。还展示了两步验证方法和嵌套(已签名和加密)JWT的处理。

jose4j / JWT Examples

JSON Web Token (JWT) Code Examples

Producing and consuming a signed JWT

And example showing simple generation and consumption of a JWT

    //
    // JSON Web Token is a compact URL-safe means of representing claims/attributes to be transferred between two parties.
    // This example demonstrates producing and consuming a signed JWT
    //

    // Generate an RSA key pair, which will be used for signing and verification of the JWT, wrapped in a JWK
    RsaJsonWebKey rsaJsonWebKey = RsaJwkGenerator.generateJwk(2048);

    // Give the JWK a Key ID (kid), which is just the polite thing to do
    rsaJsonWebKey.setKeyId("k1");

    // Create the Claims, which will be the content of the JWT
    JwtClaims claims = new JwtClaims();
    claims.setIssuer("Issuer");  // who creates the token and signs it
    claims.setAudience("Audience"); // to whom the token is intended to be sent
    claims.setExpirationTimeMinutesInTheFuture(10); // time when the token will expire (10 minutes from now)
    claims.setGeneratedJwtId(); // a unique identifier for the token
    claims.setIssuedAtToNow();  // when the token was issued/created (now)
    claims.setNotBeforeMinutesInThePast(2); // time before which the token is not yet valid (2 minutes ago)
    claims.setSubject("subject"); // the subject/principal is whom the token is about
    claims.setClaim("email","mail@example.com"); // additional claims/attributes about the subject can be added
    List<String> groups = Arrays.asList("group-one", "other-group", "group-three");
    claims.setStringListClaim("groups", groups); // multi-valued claims work too and will end up as a JSON array

    // A JWT is a JWS and/or a JWE with JSON claims as the payload.
    // In this example it is a JWS so we create a JsonWebSignature object.
    JsonWebSignature jws = new JsonWebSignature();

    // The payload of the JWS is JSON content of the JWT Claims
    jws.setPayload(claims.toJson());

    // The JWT is signed using the private key
    jws.setKey(rsaJsonWebKey.getPrivateKey());

    // Set the Key ID (kid) header because it's just the polite thing to do.
    // We only have one key in this example but a using a Key ID helps
    // facilitate a smooth key rollover process
    jws.setKeyIdHeaderValue(rsaJsonWebKey.getKeyId());

    // Set the signature algorithm on the JWT/JWS that will integrity protect the claims
    jws.setAlgorithmHeaderValue(AlgorithmIdentifiers.RSA_USING_SHA256);

    // Sign the JWS and produce the compact serialization or the complete JWT/JWS
    // representation, which is a string consisting of three dot ('.') separated
    // base64url-encoded parts in the form Header.Payload.Signature
    // If you wanted to encrypt it, you can simply set this jwt as the payload
    // of a JsonWebEncryption object and set the cty (Content Type) header to "jwt".
    String jwt = jws.getCompactSerialization();


    // Now you can do something with the JWT. Like send it to some other party
    // over the clouds and through the interwebs.
    System.out.println("JWT: " + jwt);


    // Use JwtConsumerBuilder to construct an appropriate JwtConsumer, which will
    // be used to validate and process the JWT.
    // The specific validation requirements for a JWT are context dependent, however,
    // it typically advisable to require a (reasonable) expiration time, a trusted issuer, and
    // and audience that identifies your system as the intended recipient.
    // If the JWT is encrypted too, you need only provide a decryption key or
    // decryption key resolver to the builder.
    JwtConsumer jwtConsumer = new JwtConsumerBuilder()
            .setRequireExpirationTime() // the JWT must have an expiration time
            .setMaxFutureValidityInMinutes(300) // but the  expiration time can't be too crazy
            .setAllowedClockSkewInSeconds(30) // allow some leeway in validating time based claims to account for clock skew
            .setRequireSubject() // the JWT must have a subject claim
            .setExpectedIssuer("Issuer") // whom the JWT needs to have been issued by
            .setExpectedAudience("Audience") // to whom the JWT is intended for
            .setVerificationKey(rsaJsonWebKey.getKey()) // verify the signature with the public key
            .build(); // create the JwtConsumer instance

    try
    {
        //  Validate the JWT and process it to the Claims
        JwtClaims jwtClaims = jwtConsumer.processToClaims(jwt);
        System.out.println("JWT validation succeeded! " + jwtClaims);
    }
    catch (InvalidJwtException e)
    {
        // InvalidJwtException will be thrown, if the JWT failed processing or validation in anyway.
        // Hopefully with meaningful explanations(s) about what went wrong.
        System.out.println("Invalid JWT! " + e);
    }

Using an HTTPS JWKS endpoint

If the JWT Issuer has their public keys at a HTTPS JWKS endpoint

    // In the example above we generated a key pair and used it directly for signing and verification.
    // Key exchange in the real word, however, is rarely so simple.
    // A common pattern that's emerging is for an issuer to publish its public keys
    // as a JSON Web Key Set at an HTTPS endpoint. And for the consumer of the JWT to periodically,
    // based on cache directives or known/unknown Key IDs, retrieve the keys from the host authenticated
    // and secured endpoint.

    // The HttpsJwks retrieves and caches keys from a the given HTTPS JWKS endpoint.
    // Because it retains the JWKs after fetching them, it can and should be reused
    // to improve efficiency by reducing the number of outbound calls the the endpoint.
    HttpsJwks httpsJkws = new HttpsJwks("https://example.com/jwks");

    // The HttpsJwksVerificationKeyResolver uses JWKs obtained from the HttpsJwks and will select the
    // most appropriate one to use for verification based on the Key ID and other factors provided
    // in the header of the JWS/JWT.
    HttpsJwksVerificationKeyResolver httpsJwksKeyResolver = new HttpsJwksVerificationKeyResolver(httpsJkws);


    // Use JwtConsumerBuilder to construct an appropriate JwtConsumer, which will
    // be used to validate and process the JWT. But, in this case, provide it with
    // the HttpsJwksVerificationKeyResolver instance rather than setting the
    // verification key explicitly.
    jwtConsumer = new JwtConsumerBuilder()
            // ... other set up of the JwtConsumerBuilder ...
            .setVerificationKeyResolver(httpsJwksKeyResolver)
            // ...
            .build();

Using JWKs

Or you got some JWKs out-of-band from the JWT Issuer.

    // There's also a key resolver that selects from among a given list of JWKs using the Key ID
    // and other factors provided in the header of the JWS/JWT.
    JsonWebKeySet jsonWebKeySet = new JsonWebKeySet(rsaJsonWebKey);
    JwksVerificationKeyResolver jwksResolver = new JwksVerificationKeyResolver(jsonWebKeySet.getJsonWebKeys());
    jwtConsumer = new JwtConsumerBuilder()
            // ... other set up of the JwtConsumerBuilder ...
            .setVerificationKeyResolver(jwksResolver)
            // ...
            .build();

X.509

X.509 Certificates? No problem, there's a Resolver for that too.

    // Sometimes X509 certificate(s) are provided out-of-band somehow by the signer/issuer
    // and the X509VerificationKeyResolver is helpful for that situation. It will use
    // the X.509 Certificate Thumbprint Headers (x5t or x5t#S256) from the JWS/JWT to
    // select from among the provided certificates to get the public key for verification.
    X509Util x509Util = new X509Util();
    X509Certificate certificate = x509Util.fromBase64Der(
            "MIIDQjCCAiqgAwIBAgIGATz/FuLiMA0GCSqGSIb3DQEBBQUAMGIxCzAJB" +
            "gNVBAYTAlVTMQswCQYDVQQIEwJDTzEPMA0GA1UEBxMGRGVudmVyMRwwGgYD" +
            "VQQKExNQaW5nIElkZW50aXR5IENvcnAuMRcwFQYDVQQDEw5CcmlhbiBDYW1" +
            "wYmVsbDAeFw0xMzAyMjEyMzI5MTVaFw0xODA4MTQyMjI5MTVaMGIxCzAJBg" +
            "NVBAYTAlVTMQswCQYDVQQIEwJDTzEPMA0GA1UEBxMGRGVudmVyMRwwGgYDV" +
            "QQKExNQaW5nIElkZW50aXR5IENvcnAuMRcwFQYDVQQDEw5CcmlhbiBDYW1w" +
            "YmVsbDCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAL64zn8/QnH" +
            "YMeZ0LncoXaEde1fiLm1jHjmQsF/449IYALM9if6amFtPDy2yvz3YlRij66" +
            "s5gyLCyO7ANuVRJx1NbgizcAblIgjtdf/u3WG7K+IiZhtELto/A7Fck9Ws6" +
            "SQvzRvOE8uSirYbgmj6He4iO8NCyvaK0jIQRMMGQwsU1quGmFgHIXPLfnpn" +
            "fajr1rVTAwtgV5LEZ4Iel+W1GC8ugMhyr4/p1MtcIM42EA8BzE6ZQqC7VPq" +
            "PvEjZ2dbZkaBhPbiZAS3YeYBRDWm1p1OZtWamT3cEvqqPpnjL1XyW+oyVVk" +
            "aZdklLQp2Btgt9qr21m42f4wTw+Xrp6rCKNb0CAwEAATANBgkqhkiG9w0BA" +
            "QUFAAOCAQEAh8zGlfSlcI0o3rYDPBB07aXNswb4ECNIKG0CETTUxmXl9KUL" +
            "+9gGlqCz5iWLOgWsnrcKcY0vXPG9J1r9AqBNTqNgHq2G03X09266X5CpOe1" +
            "zFo+Owb1zxtp3PehFdfQJ610CDLEaS9V9Rqp17hCyybEpOGVwe8fnk+fbEL" +
            "2Bo3UPGrpsHzUoaGpDftmWssZkhpBJKVMJyf/RuP2SmmaIzmnw9JiSlYhzo" +
            "4tpzd5rFXhjRbg4zW9C+2qok+2+qDM1iJ684gPHMIY8aLWrdgQTxkumGmTq" +
            "gawR+N5MDtdPTEQ0XfIBc2cJEUyMTY5MPvACWpkA6SdS4xSvdXK3IVfOWA==");

    X509Certificate otherCertificate = x509Util.fromBase64Der(
            "MIICUDCCAbkCBETczdcwDQYJKoZIhvcNAQEFBQAwbzELMAkGA1UEBhMCVVMxCzAJ" +
            "BgNVBAgTAkNPMQ8wDQYDVQQHEwZEZW52ZXIxFTATBgNVBAoTDFBpbmdJZGVudGl0" +
            "eTEXMBUGA1UECxMOQnJpYW4gQ2FtcGJlbGwxEjAQBgNVBAMTCWxvY2FsaG9zdDAe" +
            "Fw0wNjA4MTExODM1MDNaFw0zMzEyMjcxODM1MDNaMG8xCzAJBgNVBAYTAlVTMQsw" +
            "CQYDVQQIEwJDTzEPMA0GA1UEBxMGRGVudmVyMRUwEwYDVQQKEwxQaW5nSWRlbnRp" +
            "dHkxFzAVBgNVBAsTDkJyaWFuIENhbXBiZWxsMRIwEAYDVQQDEwlsb2NhbGhvc3Qw" +
            "gZ8wDQYJKoZIhvcNAQEBBQADgY0AMIGJAoGBAJLrpeiY/Ai2gGFxNY8Tm/QSO8qg" +
            "POGKDMAT08QMyHRlxW8fpezfBTAtKcEsztPzwYTLWmf6opfJT+5N6cJKacxWchn/" +
            "dRrzV2BoNuz1uo7wlpRqwcaOoi6yHuopNuNO1ms1vmlv3POq5qzMe6c1LRGADyZh" +
            "i0KejDX6+jVaDiUTAgMBAAEwDQYJKoZIhvcNAQEFBQADgYEAMojbPEYJiIWgQzZc" +
            "QJCQeodtKSJl5+lA8MWBBFFyZmvZ6jUYglIQdLlc8Pu6JF2j/hZEeTI87z/DOT6U" +
            "uqZA83gZcy6re4wMnZvY2kWX9CsVWDCaZhnyhjBNYfhcOf0ZychoKShaEpTQ5UAG" +
            "wvYYcbqIWC04GAZYVsZxlPl9hoA=");

    X509VerificationKeyResolver x509VerificationKeyResolver = new X509VerificationKeyResolver(certificate, otherCertificate);

    // Optionally the X509VerificationKeyResolver can attempt to verify the signature
    // with the key from each of the provided certificates, if no X.509 Certificate
    // Thumbprint Header is present in the JWT/JWS.
    x509VerificationKeyResolver.setTryAllOnNoThumbHeader(true);

    jwtConsumer = new JwtConsumerBuilder()
            // ... other set up of the JwtConsumerBuilder ...
            .setVerificationKeyResolver(x509VerificationKeyResolver)
            // ...
            .build();


    // Note that on the producing side, the X.509 Certificate Thumbprint Header
    // can be set like this on the JWS (which is the JWT)
    jws.setX509CertSha1ThumbprintHeaderValue(certificate);

Something else? (X.509 Certificates at some HTTPS endpoint, maybe)

The key resolver functionality is extensible and customizable so you can plug in your own implantation of the VerificationKeyResolver interface to do whatever you need. For example, here's a VerificationKeyResolver implementation designed to work with the "Key ID X.509 URL" that PingFederate provides in support of JWT access token validation.

Two-pass JWT consumption

Sometimes you'll need to crack open the JWT in order to know who issued it and how to validate it, which can be done efficiently and relatively easily using a two-pass consumption approach.

    // In some cases you won't have enough information to set up your JWT consumer without cracking open
    // the JWT first. For example, in some contexts you might not know who issued the token without looking
    // at the "iss" claim inside the JWT.
    // This can be done efficiently and relatively easily using two JwtConsumers in a "two-pass" validation
    // of sorts - the first JwtConsumer parses the JWT and the second one does the actual validation.

    // Build a JwtConsumer that doesn't check signatures or do any validation.
    JwtConsumer firstPassJwtConsumer = new JwtConsumerBuilder()
            .setSkipAllValidators()
            .setDisableRequireSignature()
            .setSkipSignatureVerification()
            .build();

    //The first JwtConsumer is basically just used to parse the JWT into a JwtContext object.
    JwtContext jwtContext = firstPassJwtConsumer.process(jwt);

    // From the JwtContext we can get the issuer, or whatever else we might need,
    // to lookup or figure out the kind of validation policy to apply
    String issuer = jwtContext.getJwtClaims().getIssuer();

    // Just using the same key here but you might, for example, have a JWKS URIs configured for
    // each issuer, which you'd use to set up a HttpsJwksVerificationKeyResolver
    Key verificationKey = rsaJsonWebKey.getKey();

    // Using info from the JwtContext, this JwtConsumer is set up to verify
    // the signature and validate the claims.
    JwtConsumer secondPassJwtConsumer = new JwtConsumerBuilder()
            .setExpectedIssuer(issuer)
            .setVerificationKey(verificationKey)
            .setRequireExpirationTime()
            .setAllowedClockSkewInSeconds(30)
            .setRequireSubject()
            .setExpectedAudience("Audience")
            .build();

    // Finally using the second JwtConsumer to actually validate the JWT. This operates on
    // the JwtContext from the first processing pass, which avoids redundant parsing/processing.
    secondPassJwtConsumer.processContext(jwtContext);

Producing and consuming a nested (signed and encrypted) JWT

 

    // Generate an EC key pair, which will be used for signing and verification of the JWT, wrapped in a JWK
    EllipticCurveJsonWebKey senderJwk = EcJwkGenerator.generateJwk(EllipticCurves.P256);

    // Give the JWK a Key ID (kid), which is just the polite thing to do
    senderJwk.setKeyId("sender's key");


    // Generate an EC key pair, wrapped in a JWK, which will be used for encryption and decryption of the JWT
    EllipticCurveJsonWebKey receiverJwk = EcJwkGenerator.generateJwk(EllipticCurves.P256);

    // Give the JWK a Key ID (kid), which is just the polite thing to do
    receiverJwk.setKeyId("receiver's key");

    // Create the Claims, which will be the content of the JWT
    JwtClaims claims = new JwtClaims();
    claims.setIssuer("sender");  // who creates the token and signs it
    claims.setAudience("receiver"); // to whom the token is intended to be sent
    claims.setExpirationTimeMinutesInTheFuture(10); // time when the token will expire (10 minutes from now)
    claims.setGeneratedJwtId(); // a unique identifier for the token
    claims.setIssuedAtToNow();  // when the token was issued/created (now)
    claims.setNotBeforeMinutesInThePast(2); // time before which the token is not yet valid (2 minutes ago)
    claims.setSubject("subject"); // the subject/principal is whom the token is about
    claims.setClaim("email","mail@example.com"); // additional claims/attributes about the subject can be added
    List<String> groups = Arrays.asList("group-1", "other-group", "group-3");
    claims.setStringListClaim("groups", groups); // multi-valued claims work too and will end up as a JSON array

    // A JWT is a JWS and/or a JWE with JSON claims as the payload.
    // In this example it is a JWS nested inside a JWE
    // So we first create a JsonWebSignature object.
    JsonWebSignature jws = new JsonWebSignature();

    // The payload of the JWS is JSON content of the JWT Claims
    jws.setPayload(claims.toJson());

    // The JWT is signed using the sender's private key
    jws.setKey(senderJwk.getPrivateKey());

    // Set the Key ID (kid) header because it's just the polite thing to do.
    // We only have one signing key in this example but a using a Key ID helps
    // facilitate a smooth key rollover process
    jws.setKeyIdHeaderValue(senderJwk.getKeyId());

    // Set the signature algorithm on the JWT/JWS that will integrity protect the claims
    jws.setAlgorithmHeaderValue(AlgorithmIdentifiers.ECDSA_USING_P256_CURVE_AND_SHA256);

    // Sign the JWS and produce the compact serialization, which will be the inner JWT/JWS
    // representation, which is a string consisting of three dot ('.') separated
    // base64url-encoded parts in the form Header.Payload.Signature
    String innerJwt = jws.getCompactSerialization();

    // The outer JWT is a JWE
    JsonWebEncryption jwe = new JsonWebEncryption();

    // The output of the ECDH-ES key agreement will encrypt a randomly generated content encryption key
    jwe.setAlgorithmHeaderValue(KeyManagementAlgorithmIdentifiers.ECDH_ES_A128KW);

    // The content encryption key is used to encrypt the payload
    // with a composite AES-CBC / HMAC SHA2 encryption algorithm
    String encAlg = ContentEncryptionAlgorithmIdentifiers.AES_128_CBC_HMAC_SHA_256;
    jwe.setEncryptionMethodHeaderParameter(encAlg);

    // We encrypt to the receiver using their public key
    jwe.setKey(receiverJwk.getPublicKey());
    jwe.setKeyIdHeaderValue(receiverJwk.getKeyId());

    // A nested JWT requires that the cty (Content Type) header be set to "JWT" in the outer JWT
    jwe.setContentTypeHeaderValue("JWT");

    // The inner JWT is the payload of the outer JWT
    jwe.setPayload(innerJwt);

    // Produce the JWE compact serialization, which is the complete JWT/JWE representation,
    // which is a string consisting of five dot ('.') separated
    // base64url-encoded parts in the form Header.EncryptedKey.IV.Ciphertext.AuthenticationTag
    String jwt = jwe.getCompactSerialization();


    // Now you can do something with the JWT. Like send it to some other party
    // over the clouds and through the interwebs.
    System.out.println("JWT: " + jwt);


    // Use JwtConsumerBuilder to construct an appropriate JwtConsumer, which will
    // be used to validate and process the JWT.
    // The specific validation requirements for a JWT are context dependent, however,
    // it typically advisable to require a expiration time, a trusted issuer, and
    // and audience that identifies your system as the intended recipient.
    JwtConsumer jwtConsumer = new JwtConsumerBuilder()
            .setRequireExpirationTime() // the JWT must have an expiration time
            .setRequireSubject() // the JWT must have a subject claim
            .setExpectedIssuer("sender") // whom the JWT needs to have been issued by
            .setExpectedAudience("receiver") // to whom the JWT is intended for
            .setDecryptionKey(receiverJwk.getPrivateKey()) // decrypt with the receiver's private key
            .setVerificationKey(senderJwk.getPublicKey()) // verify the signature with the sender's public key
            .build(); // create the JwtConsumer instance

    try
    {
        //  Validate the JWT and process it to the Claims
        JwtClaims jwtClaims = jwtConsumer.processToClaims(jwt);
        System.out.println("JWT validation succeeded! " + jwtClaims);
    }
    catch (InvalidJwtException e)
    {
        // InvalidJwtException will be thrown, if the JWT failed processing or validation in anyway.
        // Hopefully with meaningful explanations(s) about what went wrong.
        System.out.println("Invalid JWT! " + e);
    }

https://bitbucket.org/b_c/jose4j/wiki/JWT%20Examples#clone

 
[root@master logs]# cat hadoop-root-datanode-master.log 2025-08-23 21:26:53,162 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting DataNode STARTUP_MSG: host = master/192.168.88.8 STARTUP_MSG: args = [] STARTUP_MSG: version = 3.3.6 STARTUP_MSG: classpath = /server/hadoop/etc/hadoop:/server/hadoop/share/hadoop/common/lib/listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar:/server/hadoop/share/hadoop/common/lib/kerby-pkix-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jackson-annotations-2.12.7.jar:/server/hadoop/share/hadoop/common/lib/netty-handler-ssl-ocsp-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/metrics-core-3.2.4.jar:/server/hadoop/share/hadoop/common/lib/netty-resolver-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/commons-text-1.10.0.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/hadoop-shaded-protobuf_3_7-1.1.1.jar:/server/hadoop/share/hadoop/common/lib/jackson-mapper-asl-1.9.13.jar:/server/hadoop/share/hadoop/common/lib/jetty-server-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/netty-common-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerby-util-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/hadoop-annotations-3.3.6.jar:/server/hadoop/share/hadoop/common/lib/failureaccess-1.0.jar:/server/hadoop/share/hadoop/common/lib/jersey-json-1.20.jar:/server/hadoop/share/hadoop/common/lib/nimbus-jose-jwt-9.8.1.jar:/server/hadoop/share/hadoop/common/lib/netty-resolver-dns-native-macos-4.1.89.Final-osx-x86_64.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/snappy-java-1.1.8.2.jar:/server/hadoop/share/hadoop/common/lib/kerb-identity-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/kerb-util-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/kerb-client-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/gson-2.9.0.jar:/server/hadoop/share/hadoop/common/lib/commons-logging-1.1.3.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-native-unix-common-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-redis-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-http2-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-native-kqueue-4.1.89.Final-osx-aarch_64.jar:/server/hadoop/share/hadoop/common/lib/jetty-util-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/jsp-api-2.1.jar:/server/hadoop/share/hadoop/common/lib/jetty-util-ajax-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/zookeeper-3.6.3.jar:/server/hadoop/share/hadoop/common/lib/guava-27.0-jre.jar:/server/hadoop/share/hadoop/common/lib/httpcore-4.4.13.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-rxtx-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerb-admin-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/curator-client-5.2.0.jar:/server/hadoop/share/hadoop/common/lib/netty-buffer-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-native-epoll-4.1.89.Final-linux-x86_64.jar:/server/hadoop/share/hadoop/common/lib/jcip-annotations-1.0-1.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-xml-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/javax.servlet-api-3.1.0.jar:/server/hadoop/share/hadoop/common/lib/netty-resolver-dns-classes-macos-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/curator-recipes-5.2.0.jar:/server/hadoop/share/hadoop/common/lib/commons-net-3.9.0.jar:/server/hadoop/share/hadoop/common/lib/jackson-databind-2.12.7.1.jar:/server/hadoop/share/hadoop/common/lib/commons-beanutils-1.9.4.jar:/server/hadoop/share/hadoop/common/lib/netty-resolver-dns-native-macos-4.1.89.Final-osx-aarch_64.jar:/server/hadoop/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/server/hadoop/share/hadoop/common/lib/jettison-1.5.4.jar:/server/hadoop/share/hadoop/common/lib/slf4j-api-1.7.36.jar:/server/hadoop/share/hadoop/common/lib/jsr305-3.0.2.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-dns-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/animal-sniffer-annotations-1.17.jar:/server/hadoop/share/hadoop/common/lib/kerb-core-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jetty-servlet-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-socks-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/commons-io-2.8.0.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-classes-epoll-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/re2j-1.1.jar:/server/hadoop/share/hadoop/common/lib/netty-handler-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/commons-daemon-1.0.13.jar:/server/hadoop/share/hadoop/common/lib/slf4j-reload4j-1.7.36.jar:/server/hadoop/share/hadoop/common/lib/commons-codec-1.15.jar:/server/hadoop/share/hadoop/common/lib/woodstox-core-5.4.0.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-stomp-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/zookeeper-jute-3.6.3.jar:/server/hadoop/share/hadoop/common/lib/kerby-asn1-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/kerby-config-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jackson-core-asl-1.9.13.jar:/server/hadoop/share/hadoop/common/lib/stax2-api-4.2.1.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-http-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerb-crypto-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jakarta.activation-api-1.2.1.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-classes-kqueue-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/server/hadoop/share/hadoop/common/lib/hadoop-shaded-guava-1.1.1.jar:/server/hadoop/share/hadoop/common/lib/jackson-core-2.12.7.jar:/server/hadoop/share/hadoop/common/lib/commons-configuration2-2.8.0.jar:/server/hadoop/share/hadoop/common/lib/jaxb-api-2.2.11.jar:/server/hadoop/share/hadoop/common/lib/jetty-security-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/token-provider-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/reload4j-1.2.22.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-haproxy-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/curator-framework-5.2.0.jar:/server/hadoop/share/hadoop/common/lib/jsch-0.1.55.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-native-kqueue-4.1.89.Final-osx-x86_64.jar:/server/hadoop/share/hadoop/common/lib/checker-qual-2.5.2.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-memcache-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/jetty-http-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/netty-handler-proxy-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-native-epoll-4.1.89.Final-linux-aarch_64.jar:/server/hadoop/share/hadoop/common/lib/jetty-webapp-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/jersey-server-1.19.4.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-sctp-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-resolver-dns-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerby-xdr-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jul-to-slf4j-1.7.36.jar:/server/hadoop/share/hadoop/common/lib/commons-lang3-3.12.0.jar:/server/hadoop/share/hadoop/common/lib/kerb-common-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jetty-io-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-udt-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerb-simplekdc-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jetty-xml-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/paranamer-2.3.jar:/server/hadoop/share/hadoop/common/lib/commons-math3-3.1.1.jar:/server/hadoop/share/hadoop/common/lib/audience-annotations-0.5.0.jar:/server/hadoop/share/hadoop/common/lib/jersey-servlet-1.19.4.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-smtp-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/dnsjava-2.1.7.jar:/server/hadoop/share/hadoop/common/lib/commons-collections-3.2.2.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-mqtt-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerb-server-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/httpclient-4.5.13.jar:/server/hadoop/share/hadoop/common/lib/netty-all-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/j2objc-annotations-1.1.jar:/server/hadoop/share/hadoop/common/lib/jersey-core-1.19.4.jar:/server/hadoop/share/hadoop/common/lib/commons-cli-1.2.jar:/server/hadoop/share/hadoop/common/lib/hadoop-auth-3.3.6.jar:/server/hadoop/share/hadoop/common/lib/jsr311-api-1.1.1.jar:/server/hadoop/share/hadoop/common/lib/commons-compress-1.21.jar:/server/hadoop/share/hadoop/common/lib/avro-1.7.7.jar:/server/hadoop/share/hadoop/common/hadoop-registry-3.3.6.jar:/server/hadoop/share/hadoop/common/hadoop-common-3.3.6.jar:/server/hadoop/share/hadoop/common/hadoop-kms-3.3.6.jar:/server/hadoop/share/hadoop/common/hadoop-nfs-3.3.6.jar:/server/hadoop/share/hadoop/common/hadoop-common-3.3.6-tests.jar:/server/hadoop/share/hadoop/hdfs:/server/hadoop/share/hadoop/hdfs/lib/listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar:/server/hadoop/share/hadoop/hdfs/lib/kerby-pkix-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jackson-annotations-2.12.7.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-handler-ssl-ocsp-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/metrics-core-3.2.4.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-resolver-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-text-1.10.0.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/hadoop-shaded-protobuf_3_7-1.1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jackson-mapper-asl-1.9.13.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-server-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-common-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerby-util-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/hadoop-annotations-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/lib/failureaccess-1.0.jar:/server/hadoop/share/hadoop/hdfs/lib/jersey-json-1.20.jar:/server/hadoop/share/hadoop/hdfs/lib/nimbus-jose-jwt-9.8.1.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-resolver-dns-native-macos-4.1.89.Final-osx-x86_64.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/snappy-java-1.1.8.2.jar:/server/hadoop/share/hadoop/hdfs/lib/okio-2.8.0.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-identity-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-util-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-client-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/gson-2.9.0.jar:/server/hadoop/share/hadoop/hdfs/lib/kotlin-stdlib-1.4.10.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-native-unix-common-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-redis-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-http2-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-native-kqueue-4.1.89.Final-osx-aarch_64.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-util-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-util-ajax-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/zookeeper-3.6.3.jar:/server/hadoop/share/hadoop/hdfs/lib/guava-27.0-jre.jar:/server/hadoop/share/hadoop/hdfs/lib/httpcore-4.4.13.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-rxtx-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-admin-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/curator-client-5.2.0.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-buffer-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-native-epoll-4.1.89.Final-linux-x86_64.jar:/server/hadoop/share/hadoop/hdfs/lib/jcip-annotations-1.0-1.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-xml-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/javax.servlet-api-3.1.0.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-resolver-dns-classes-macos-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/curator-recipes-5.2.0.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-net-3.9.0.jar:/server/hadoop/share/hadoop/hdfs/lib/jackson-databind-2.12.7.1.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-beanutils-1.9.4.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-resolver-dns-native-macos-4.1.89.Final-osx-aarch_64.jar:/server/hadoop/share/hadoop/hdfs/lib/jaxb-impl-2.2.3-1.jar:/server/hadoop/share/hadoop/hdfs/lib/jettison-1.5.4.jar:/server/hadoop/share/hadoop/hdfs/lib/jsr305-3.0.2.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-dns-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/animal-sniffer-annotations-1.17.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-core-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-servlet-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-socks-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-io-2.8.0.jar:/server/hadoop/share/hadoop/hdfs/lib/leveldbjni-all-1.8.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-classes-epoll-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/re2j-1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-handler-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-codec-1.15.jar:/server/hadoop/share/hadoop/hdfs/lib/woodstox-core-5.4.0.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-stomp-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/zookeeper-jute-3.6.3.jar:/server/hadoop/share/hadoop/hdfs/lib/kerby-asn1-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/kerby-config-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jackson-core-asl-1.9.13.jar:/server/hadoop/share/hadoop/hdfs/lib/stax2-api-4.2.1.jar:/server/hadoop/share/hadoop/hdfs/lib/kotlin-stdlib-common-1.4.10.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-http-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-crypto-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-3.10.6.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/jakarta.activation-api-1.2.1.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-classes-kqueue-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/server/hadoop/share/hadoop/hdfs/lib/hadoop-shaded-guava-1.1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jackson-core-2.12.7.jar:/server/hadoop/share/hadoop/hdfs/lib/okhttp-4.9.3.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-configuration2-2.8.0.jar:/server/hadoop/share/hadoop/hdfs/lib/jaxb-api-2.2.11.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-security-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/token-provider-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/reload4j-1.2.22.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-haproxy-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/curator-framework-5.2.0.jar:/server/hadoop/share/hadoop/hdfs/lib/jsch-0.1.55.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-native-kqueue-4.1.89.Final-osx-x86_64.jar:/server/hadoop/share/hadoop/hdfs/lib/checker-qual-2.5.2.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-memcache-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-http-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-handler-proxy-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-native-epoll-4.1.89.Final-linux-aarch_64.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-webapp-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/jersey-server-1.19.4.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-sctp-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-resolver-dns-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerby-xdr-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/HikariCP-java7-2.4.12.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-lang3-3.12.0.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-common-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-io-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-udt-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-simplekdc-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-xml-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/paranamer-2.3.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-math3-3.1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/audience-annotations-0.5.0.jar:/server/hadoop/share/hadoop/hdfs/lib/jersey-servlet-1.19.4.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-smtp-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/dnsjava-2.1.7.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-collections-3.2.2.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-mqtt-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-server-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/json-simple-1.1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/httpclient-4.5.13.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-all-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/j2objc-annotations-1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jersey-core-1.19.4.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/server/hadoop/share/hadoop/hdfs/lib/hadoop-auth-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/lib/jsr311-api-1.1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-compress-1.21.jar:/server/hadoop/share/hadoop/hdfs/lib/avro-1.7.7.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-rbf-3.3.6-tests.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-nfs-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-native-client-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-native-client-3.3.6-tests.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-client-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-rbf-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-httpfs-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-3.3.6-tests.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-client-3.3.6-tests.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-common-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-3.3.6-tests.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-uploader-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-nativetask-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-examples-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-app-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-core-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-3.3.6.jar:/server/hadoop/share/hadoop/yarn:/server/hadoop/share/hadoop/yarn/lib/snakeyaml-2.0.jar:/server/hadoop/share/hadoop/yarn/lib/bcpkix-jdk15on-1.68.jar:/server/hadoop/share/hadoop/yarn/lib/jackson-jaxrs-json-provider-2.12.7.jar:/server/hadoop/share/hadoop/yarn/lib/jetty-client-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jersey-client-1.19.4.jar:/server/hadoop/share/hadoop/yarn/lib/javax-websocket-client-impl-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jline-3.9.0.jar:/server/hadoop/share/hadoop/yarn/lib/javax.websocket-client-api-1.0.jar:/server/hadoop/share/hadoop/yarn/lib/swagger-annotations-1.5.4.jar:/server/hadoop/share/hadoop/yarn/lib/mssql-jdbc-6.2.1.jre7.jar:/server/hadoop/share/hadoop/yarn/lib/guice-servlet-4.0.jar:/server/hadoop/share/hadoop/yarn/lib/jetty-annotations-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jetty-jndi-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/aopalliance-1.0.jar:/server/hadoop/share/hadoop/yarn/lib/jackson-jaxrs-base-2.12.7.jar:/server/hadoop/share/hadoop/yarn/lib/json-io-2.5.1.jar:/server/hadoop/share/hadoop/yarn/lib/javax.inject-1.jar:/server/hadoop/share/hadoop/yarn/lib/objenesis-2.6.jar:/server/hadoop/share/hadoop/yarn/lib/asm-tree-9.4.jar:/server/hadoop/share/hadoop/yarn/lib/geronimo-jcache_1.0_spec-1.0-alpha-1.jar:/server/hadoop/share/hadoop/yarn/lib/websocket-common-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jetty-plus-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jna-5.2.0.jar:/server/hadoop/share/hadoop/yarn/lib/javax.websocket-api-1.0.jar:/server/hadoop/share/hadoop/yarn/lib/javax-websocket-server-impl-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/websocket-api-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/websocket-client-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jackson-module-jaxb-annotations-2.12.7.jar:/server/hadoop/share/hadoop/yarn/lib/bcprov-jdk15on-1.68.jar:/server/hadoop/share/hadoop/yarn/lib/asm-commons-9.4.jar:/server/hadoop/share/hadoop/yarn/lib/guice-4.0.jar:/server/hadoop/share/hadoop/yarn/lib/websocket-servlet-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/websocket-server-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jakarta.xml.bind-api-2.3.2.jar:/server/hadoop/share/hadoop/yarn/lib/fst-2.50.jar:/server/hadoop/share/hadoop/yarn/lib/jersey-guice-1.19.4.jar:/server/hadoop/share/hadoop/yarn/lib/ehcache-3.3.1.jar:/server/hadoop/share/hadoop/yarn/lib/java-util-1.9.0.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-registry-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-nodemanager-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-sharedcachemanager-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-common-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-common-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-web-proxy-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-timeline-pluginstorage-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-tests-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-client-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-applications-mawo-core-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-router-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-services-api-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-api-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-services-core-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-3.3.6.jar STARTUP_MSG: build = https://github.com/apache/hadoop.git -r 1be78238728da9266a4f88195058f08fd012bf9c; compiled by 'ubuntu' on 2023-06-18T08:22Z STARTUP_MSG: java = 1.8.0_401 ************************************************************/ 2025-08-23 21:26:53,175 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: registered UNIX signal handlers for [TERM, HUP, INT] 2025-08-23 21:26:53,573 INFO org.apache.hadoop.hdfs.server.datanode.checker.ThrottledAsyncChecker: Scheduling a check for [DISK]file:/server/hadoop/data/dn 2025-08-23 21:26:53,762 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: Loaded properties from hadoop-metrics2.properties 2025-08-23 21:26:53,865 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled Metric snapshot period at 10 second(s). 2025-08-23 21:26:53,865 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system started 2025-08-23 21:26:54,362 INFO org.apache.hadoop.hdfs.server.common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2025-08-23 21:26:54,408 INFO org.apache.hadoop.hdfs.server.datanode.BlockScanner: Initialized block scanner with targetBytesPerSec 1048576 2025-08-23 21:26:54,417 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Configured hostname is master 2025-08-23 21:26:54,417 INFO org.apache.hadoop.hdfs.server.common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2025-08-23 21:26:54,439 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting DataNode with maxLockedMemory = 0 2025-08-23 21:26:54,484 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened streaming server at /0.0.0.0:9866 2025-08-23 21:26:54,486 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwidth is 104857600 bytes/s 2025-08-23 21:26:54,486 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Number threads for balancing is 100 2025-08-23 21:26:54,598 INFO org.eclipse.jetty.util.log: Logging initialized @2208ms to org.eclipse.jetty.util.log.Slf4jLog 2025-08-23 21:26:54,928 WARN org.apache.hadoop.security.authentication.server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /root/hadoop-http-auth-signature-secret 2025-08-23 21:26:54,939 INFO org.apache.hadoop.http.HttpRequestLog: Http request log for http.requests.datanode is not defined 2025-08-23 21:26:54,950 INFO org.apache.hadoop.http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) 2025-08-23 21:26:54,955 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context datanode 2025-08-23 21:26:54,955 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2025-08-23 21:26:54,955 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2025-08-23 21:26:55,012 INFO org.apache.hadoop.http.HttpServer2: Jetty bound to port 43231 2025-08-23 21:26:55,014 INFO org.eclipse.jetty.server.Server: jetty-9.4.51.v20230217; built: 2023-02-17T08:19:37.309Z; git: b45c405e4544384de066f814ed42ae3dceacdd49; jvm 1.8.0_401-b10 2025-08-23 21:26:55,087 INFO org.eclipse.jetty.server.session: DefaultSessionIdManager workerName=node0 2025-08-23 21:26:55,087 INFO org.eclipse.jetty.server.session: No SessionScavenger set, using defaults 2025-08-23 21:26:55,090 INFO org.eclipse.jetty.server.session: node0 Scavenging every 600000ms 2025-08-23 21:26:55,121 INFO org.eclipse.jetty.server.handler.ContextHandler: Started o.e.j.s.ServletContextHandler@4fad9bb2{logs,/logs,file:///server/hadoop/logs/,AVAILABLE} 2025-08-23 21:26:55,122 INFO org.eclipse.jetty.server.handler.ContextHandler: Started o.e.j.s.ServletContextHandler@1941a8ff{static,/static,file:///server/hadoop/share/hadoop/hdfs/webapps/static/,AVAILABLE} 2025-08-23 21:26:55,320 INFO org.eclipse.jetty.server.handler.ContextHandler: Started o.e.j.w.WebAppContext@35390ee3{datanode,/,file:///server/hadoop/share/hadoop/hdfs/webapps/datanode/,AVAILABLE}{file:/server/hadoop/share/hadoop/hdfs/webapps/datanode} 2025-08-23 21:26:55,337 INFO org.eclipse.jetty.server.AbstractConnector: Started ServerConnector@66565121{HTTP/1.1, (http/1.1)}{localhost:43231} 2025-08-23 21:26:55,337 INFO org.eclipse.jetty.server.Server: Started @2947ms 2025-08-23 21:26:55,504 WARN org.apache.hadoop.hdfs.server.datanode.web.DatanodeHttpServer: Got null for restCsrfPreventionFilter - will not do any filtering. 2025-08-23 21:26:55,620 INFO org.apache.hadoop.hdfs.server.datanode.web.DatanodeHttpServer: Listening HTTP traffic on /0.0.0.0:9864 2025-08-23 21:26:55,636 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: dnUserName = root 2025-08-23 21:26:55,637 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: supergroup = supergroup 2025-08-23 21:26:55,638 INFO org.apache.hadoop.util.JvmPauseMonitor: Starting JVM pause monitor 2025-08-23 21:26:55,857 INFO org.apache.hadoop.ipc.CallQueueManager: Using callQueue: class java.util.concurrent.LinkedBlockingQueue, queueCapacity: 1000, scheduler: class org.apache.hadoop.ipc.DefaultRpcScheduler, ipcBackoff: false. 2025-08-23 21:26:55,878 INFO org.apache.hadoop.ipc.Server: Listener at 0.0.0.0:9867 2025-08-23 21:26:55,883 INFO org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 9867 2025-08-23 21:26:56,440 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened IPC server at /0.0.0.0:9867 2025-08-23 21:26:56,495 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Refresh request received for nameservices: null 2025-08-23 21:26:56,507 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: Unable to get NameNode addresses. java.io.IOException: Incorrect configuration: namenode address dfs.namenode.servicerpc-address.[] or dfs.namenode.rpc-address.[] is not configured. at org.apache.hadoop.hdfs.DFSUtil.getNNServiceRpcAddressesForCluster(DFSUtil.java:656) at org.apache.hadoop.hdfs.server.datanode.BlockPoolManager.refreshNamenodes(BlockPoolManager.java:157) at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1755) at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:564) at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:3148) at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:3054) at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:3098) at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:3242) at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:3266) 2025-08-23 21:26:56,522 INFO org.eclipse.jetty.server.handler.ContextHandler: Stopped o.e.j.w.WebAppContext@35390ee3{datanode,/,null,STOPPED}{file:/server/hadoop/share/hadoop/hdfs/webapps/datanode} 2025-08-23 21:26:56,527 INFO org.eclipse.jetty.server.AbstractConnector: Stopped ServerConnector@66565121{HTTP/1.1, (http/1.1)}{localhost:0} 2025-08-23 21:26:56,527 INFO org.eclipse.jetty.server.session: node0 Stopped scavenging 2025-08-23 21:26:56,527 INFO org.eclipse.jetty.server.handler.ContextHandler: Stopped o.e.j.s.ServletContextHandler@1941a8ff{static,/static,file:///server/hadoop/share/hadoop/hdfs/webapps/static/,STOPPED} 2025-08-23 21:26:56,528 INFO org.eclipse.jetty.server.handler.ContextHandler: Stopped o.e.j.s.ServletContextHandler@4fad9bb2{logs,/logs,file:///server/hadoop/logs/,STOPPED} 2025-08-23 21:26:56,537 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Waiting up to 30 seconds for transfer threads to complete 2025-08-23 21:26:56,537 INFO org.apache.hadoop.ipc.Server: Stopping server on 9867 2025-08-23 21:26:56,542 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping DataNode metrics system... 2025-08-23 21:26:56,542 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system stopped. 2025-08-23 21:26:56,543 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system shutdown complete. 2025-08-23 21:26:56,548 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Shutdown complete. 2025-08-23 21:26:56,548 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: Exception in secureMain java.io.IOException: No services to connect, missing NameNode address. at org.apache.hadoop.hdfs.server.datanode.BlockPoolManager.refreshNamenodes(BlockPoolManager.java:165) at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1755) at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:564) at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:3148) at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:3054) at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:3098) at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:3242) at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:3266) 2025-08-23 21:26:56,550 INFO org.apache.hadoop.util.ExitUtil: Exiting with status 1: java.io.IOException: No services to connect, missing NameNode address. 2025-08-23 21:26:56,558 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down DataNode at master/192.168.88.8 ************************************************************/ 2025-08-23 21:29:44,258 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting DataNode STARTUP_MSG: host = master/192.168.88.8 STARTUP_MSG: args = [] STARTUP_MSG: version = 3.3.6 STARTUP_MSG: classpath = /server/hadoop/etc/hadoop:/server/hadoop/share/hadoop/common/lib/listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar:/server/hadoop/share/hadoop/common/lib/kerby-pkix-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jackson-annotations-2.12.7.jar:/server/hadoop/share/hadoop/common/lib/netty-handler-ssl-ocsp-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/metrics-core-3.2.4.jar:/server/hadoop/share/hadoop/common/lib/netty-resolver-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/commons-text-1.10.0.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/hadoop-shaded-protobuf_3_7-1.1.1.jar:/server/hadoop/share/hadoop/common/lib/jackson-mapper-asl-1.9.13.jar:/server/hadoop/share/hadoop/common/lib/jetty-server-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/netty-common-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerby-util-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/hadoop-annotations-3.3.6.jar:/server/hadoop/share/hadoop/common/lib/failureaccess-1.0.jar:/server/hadoop/share/hadoop/common/lib/jersey-json-1.20.jar:/server/hadoop/share/hadoop/common/lib/nimbus-jose-jwt-9.8.1.jar:/server/hadoop/share/hadoop/common/lib/netty-resolver-dns-native-macos-4.1.89.Final-osx-x86_64.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/snappy-java-1.1.8.2.jar:/server/hadoop/share/hadoop/common/lib/kerb-identity-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/kerb-util-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/kerb-client-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/gson-2.9.0.jar:/server/hadoop/share/hadoop/common/lib/commons-logging-1.1.3.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-native-unix-common-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-redis-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-http2-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-native-kqueue-4.1.89.Final-osx-aarch_64.jar:/server/hadoop/share/hadoop/common/lib/jetty-util-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/jsp-api-2.1.jar:/server/hadoop/share/hadoop/common/lib/jetty-util-ajax-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/zookeeper-3.6.3.jar:/server/hadoop/share/hadoop/common/lib/guava-27.0-jre.jar:/server/hadoop/share/hadoop/common/lib/httpcore-4.4.13.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-rxtx-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerb-admin-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/curator-client-5.2.0.jar:/server/hadoop/share/hadoop/common/lib/netty-buffer-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-native-epoll-4.1.89.Final-linux-x86_64.jar:/server/hadoop/share/hadoop/common/lib/jcip-annotations-1.0-1.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-xml-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/javax.servlet-api-3.1.0.jar:/server/hadoop/share/hadoop/common/lib/netty-resolver-dns-classes-macos-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/curator-recipes-5.2.0.jar:/server/hadoop/share/hadoop/common/lib/commons-net-3.9.0.jar:/server/hadoop/share/hadoop/common/lib/jackson-databind-2.12.7.1.jar:/server/hadoop/share/hadoop/common/lib/commons-beanutils-1.9.4.jar:/server/hadoop/share/hadoop/common/lib/netty-resolver-dns-native-macos-4.1.89.Final-osx-aarch_64.jar:/server/hadoop/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/server/hadoop/share/hadoop/common/lib/jettison-1.5.4.jar:/server/hadoop/share/hadoop/common/lib/slf4j-api-1.7.36.jar:/server/hadoop/share/hadoop/common/lib/jsr305-3.0.2.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-dns-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/animal-sniffer-annotations-1.17.jar:/server/hadoop/share/hadoop/common/lib/kerb-core-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jetty-servlet-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-socks-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/commons-io-2.8.0.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-classes-epoll-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/re2j-1.1.jar:/server/hadoop/share/hadoop/common/lib/netty-handler-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/commons-daemon-1.0.13.jar:/server/hadoop/share/hadoop/common/lib/slf4j-reload4j-1.7.36.jar:/server/hadoop/share/hadoop/common/lib/commons-codec-1.15.jar:/server/hadoop/share/hadoop/common/lib/woodstox-core-5.4.0.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-stomp-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/zookeeper-jute-3.6.3.jar:/server/hadoop/share/hadoop/common/lib/kerby-asn1-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/kerby-config-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jackson-core-asl-1.9.13.jar:/server/hadoop/share/hadoop/common/lib/stax2-api-4.2.1.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-http-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerb-crypto-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jakarta.activation-api-1.2.1.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-classes-kqueue-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/server/hadoop/share/hadoop/common/lib/hadoop-shaded-guava-1.1.1.jar:/server/hadoop/share/hadoop/common/lib/jackson-core-2.12.7.jar:/server/hadoop/share/hadoop/common/lib/commons-configuration2-2.8.0.jar:/server/hadoop/share/hadoop/common/lib/jaxb-api-2.2.11.jar:/server/hadoop/share/hadoop/common/lib/jetty-security-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/token-provider-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/reload4j-1.2.22.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-haproxy-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/curator-framework-5.2.0.jar:/server/hadoop/share/hadoop/common/lib/jsch-0.1.55.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-native-kqueue-4.1.89.Final-osx-x86_64.jar:/server/hadoop/share/hadoop/common/lib/checker-qual-2.5.2.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-memcache-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/jetty-http-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/netty-handler-proxy-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-native-epoll-4.1.89.Final-linux-aarch_64.jar:/server/hadoop/share/hadoop/common/lib/jetty-webapp-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/jersey-server-1.19.4.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-sctp-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/netty-resolver-dns-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerby-xdr-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jul-to-slf4j-1.7.36.jar:/server/hadoop/share/hadoop/common/lib/commons-lang3-3.12.0.jar:/server/hadoop/share/hadoop/common/lib/kerb-common-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jetty-io-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/netty-transport-udt-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerb-simplekdc-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/jetty-xml-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/common/lib/paranamer-2.3.jar:/server/hadoop/share/hadoop/common/lib/commons-math3-3.1.1.jar:/server/hadoop/share/hadoop/common/lib/audience-annotations-0.5.0.jar:/server/hadoop/share/hadoop/common/lib/jersey-servlet-1.19.4.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-smtp-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/dnsjava-2.1.7.jar:/server/hadoop/share/hadoop/common/lib/commons-collections-3.2.2.jar:/server/hadoop/share/hadoop/common/lib/netty-codec-mqtt-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/kerb-server-1.0.1.jar:/server/hadoop/share/hadoop/common/lib/httpclient-4.5.13.jar:/server/hadoop/share/hadoop/common/lib/netty-all-4.1.89.Final.jar:/server/hadoop/share/hadoop/common/lib/j2objc-annotations-1.1.jar:/server/hadoop/share/hadoop/common/lib/jersey-core-1.19.4.jar:/server/hadoop/share/hadoop/common/lib/commons-cli-1.2.jar:/server/hadoop/share/hadoop/common/lib/hadoop-auth-3.3.6.jar:/server/hadoop/share/hadoop/common/lib/jsr311-api-1.1.1.jar:/server/hadoop/share/hadoop/common/lib/commons-compress-1.21.jar:/server/hadoop/share/hadoop/common/lib/avro-1.7.7.jar:/server/hadoop/share/hadoop/common/hadoop-registry-3.3.6.jar:/server/hadoop/share/hadoop/common/hadoop-common-3.3.6.jar:/server/hadoop/share/hadoop/common/hadoop-kms-3.3.6.jar:/server/hadoop/share/hadoop/common/hadoop-nfs-3.3.6.jar:/server/hadoop/share/hadoop/common/hadoop-common-3.3.6-tests.jar:/server/hadoop/share/hadoop/hdfs:/server/hadoop/share/hadoop/hdfs/lib/listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar:/server/hadoop/share/hadoop/hdfs/lib/kerby-pkix-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jackson-annotations-2.12.7.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-handler-ssl-ocsp-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/metrics-core-3.2.4.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-resolver-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-text-1.10.0.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/hadoop-shaded-protobuf_3_7-1.1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jackson-mapper-asl-1.9.13.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-server-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-common-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerby-util-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/hadoop-annotations-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/lib/failureaccess-1.0.jar:/server/hadoop/share/hadoop/hdfs/lib/jersey-json-1.20.jar:/server/hadoop/share/hadoop/hdfs/lib/nimbus-jose-jwt-9.8.1.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-resolver-dns-native-macos-4.1.89.Final-osx-x86_64.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/snappy-java-1.1.8.2.jar:/server/hadoop/share/hadoop/hdfs/lib/okio-2.8.0.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-identity-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-util-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-client-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/gson-2.9.0.jar:/server/hadoop/share/hadoop/hdfs/lib/kotlin-stdlib-1.4.10.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-native-unix-common-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-redis-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-http2-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-native-kqueue-4.1.89.Final-osx-aarch_64.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-util-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-util-ajax-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/zookeeper-3.6.3.jar:/server/hadoop/share/hadoop/hdfs/lib/guava-27.0-jre.jar:/server/hadoop/share/hadoop/hdfs/lib/httpcore-4.4.13.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-rxtx-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-admin-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/curator-client-5.2.0.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-buffer-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-native-epoll-4.1.89.Final-linux-x86_64.jar:/server/hadoop/share/hadoop/hdfs/lib/jcip-annotations-1.0-1.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-xml-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/javax.servlet-api-3.1.0.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-resolver-dns-classes-macos-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/curator-recipes-5.2.0.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-net-3.9.0.jar:/server/hadoop/share/hadoop/hdfs/lib/jackson-databind-2.12.7.1.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-beanutils-1.9.4.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-resolver-dns-native-macos-4.1.89.Final-osx-aarch_64.jar:/server/hadoop/share/hadoop/hdfs/lib/jaxb-impl-2.2.3-1.jar:/server/hadoop/share/hadoop/hdfs/lib/jettison-1.5.4.jar:/server/hadoop/share/hadoop/hdfs/lib/jsr305-3.0.2.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-dns-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/animal-sniffer-annotations-1.17.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-core-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-servlet-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-socks-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-io-2.8.0.jar:/server/hadoop/share/hadoop/hdfs/lib/leveldbjni-all-1.8.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-classes-epoll-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/re2j-1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-handler-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-codec-1.15.jar:/server/hadoop/share/hadoop/hdfs/lib/woodstox-core-5.4.0.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-stomp-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/zookeeper-jute-3.6.3.jar:/server/hadoop/share/hadoop/hdfs/lib/kerby-asn1-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/kerby-config-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jackson-core-asl-1.9.13.jar:/server/hadoop/share/hadoop/hdfs/lib/stax2-api-4.2.1.jar:/server/hadoop/share/hadoop/hdfs/lib/kotlin-stdlib-common-1.4.10.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-http-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-crypto-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-3.10.6.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/jakarta.activation-api-1.2.1.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-classes-kqueue-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/server/hadoop/share/hadoop/hdfs/lib/hadoop-shaded-guava-1.1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jackson-core-2.12.7.jar:/server/hadoop/share/hadoop/hdfs/lib/okhttp-4.9.3.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-configuration2-2.8.0.jar:/server/hadoop/share/hadoop/hdfs/lib/jaxb-api-2.2.11.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-security-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/token-provider-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/reload4j-1.2.22.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-haproxy-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/curator-framework-5.2.0.jar:/server/hadoop/share/hadoop/hdfs/lib/jsch-0.1.55.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-native-kqueue-4.1.89.Final-osx-x86_64.jar:/server/hadoop/share/hadoop/hdfs/lib/checker-qual-2.5.2.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-memcache-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-http-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-handler-proxy-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-native-epoll-4.1.89.Final-linux-aarch_64.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-webapp-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/jersey-server-1.19.4.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-sctp-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-resolver-dns-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerby-xdr-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/HikariCP-java7-2.4.12.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-lang3-3.12.0.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-common-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-io-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-transport-udt-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-simplekdc-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jetty-xml-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/hdfs/lib/paranamer-2.3.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-math3-3.1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/audience-annotations-0.5.0.jar:/server/hadoop/share/hadoop/hdfs/lib/jersey-servlet-1.19.4.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-smtp-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/dnsjava-2.1.7.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-collections-3.2.2.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-codec-mqtt-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/kerb-server-1.0.1.jar:/server/hadoop/share/hadoop/hdfs/lib/json-simple-1.1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/httpclient-4.5.13.jar:/server/hadoop/share/hadoop/hdfs/lib/netty-all-4.1.89.Final.jar:/server/hadoop/share/hadoop/hdfs/lib/j2objc-annotations-1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/jersey-core-1.19.4.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/server/hadoop/share/hadoop/hdfs/lib/hadoop-auth-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/lib/jsr311-api-1.1.1.jar:/server/hadoop/share/hadoop/hdfs/lib/commons-compress-1.21.jar:/server/hadoop/share/hadoop/hdfs/lib/avro-1.7.7.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-rbf-3.3.6-tests.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-nfs-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-native-client-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-native-client-3.3.6-tests.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-client-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-rbf-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-httpfs-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-3.3.6.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-3.3.6-tests.jar:/server/hadoop/share/hadoop/hdfs/hadoop-hdfs-client-3.3.6-tests.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-common-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-3.3.6-tests.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-uploader-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-nativetask-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-examples-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-app-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-core-3.3.6.jar:/server/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-3.3.6.jar:/server/hadoop/share/hadoop/yarn:/server/hadoop/share/hadoop/yarn/lib/snakeyaml-2.0.jar:/server/hadoop/share/hadoop/yarn/lib/bcpkix-jdk15on-1.68.jar:/server/hadoop/share/hadoop/yarn/lib/jackson-jaxrs-json-provider-2.12.7.jar:/server/hadoop/share/hadoop/yarn/lib/jetty-client-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jersey-client-1.19.4.jar:/server/hadoop/share/hadoop/yarn/lib/javax-websocket-client-impl-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jline-3.9.0.jar:/server/hadoop/share/hadoop/yarn/lib/javax.websocket-client-api-1.0.jar:/server/hadoop/share/hadoop/yarn/lib/swagger-annotations-1.5.4.jar:/server/hadoop/share/hadoop/yarn/lib/mssql-jdbc-6.2.1.jre7.jar:/server/hadoop/share/hadoop/yarn/lib/guice-servlet-4.0.jar:/server/hadoop/share/hadoop/yarn/lib/jetty-annotations-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jetty-jndi-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/aopalliance-1.0.jar:/server/hadoop/share/hadoop/yarn/lib/jackson-jaxrs-base-2.12.7.jar:/server/hadoop/share/hadoop/yarn/lib/json-io-2.5.1.jar:/server/hadoop/share/hadoop/yarn/lib/javax.inject-1.jar:/server/hadoop/share/hadoop/yarn/lib/objenesis-2.6.jar:/server/hadoop/share/hadoop/yarn/lib/asm-tree-9.4.jar:/server/hadoop/share/hadoop/yarn/lib/geronimo-jcache_1.0_spec-1.0-alpha-1.jar:/server/hadoop/share/hadoop/yarn/lib/websocket-common-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jetty-plus-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jna-5.2.0.jar:/server/hadoop/share/hadoop/yarn/lib/javax.websocket-api-1.0.jar:/server/hadoop/share/hadoop/yarn/lib/javax-websocket-server-impl-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/websocket-api-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/websocket-client-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jackson-module-jaxb-annotations-2.12.7.jar:/server/hadoop/share/hadoop/yarn/lib/bcprov-jdk15on-1.68.jar:/server/hadoop/share/hadoop/yarn/lib/asm-commons-9.4.jar:/server/hadoop/share/hadoop/yarn/lib/guice-4.0.jar:/server/hadoop/share/hadoop/yarn/lib/websocket-servlet-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/websocket-server-9.4.51.v20230217.jar:/server/hadoop/share/hadoop/yarn/lib/jakarta.xml.bind-api-2.3.2.jar:/server/hadoop/share/hadoop/yarn/lib/fst-2.50.jar:/server/hadoop/share/hadoop/yarn/lib/jersey-guice-1.19.4.jar:/server/hadoop/share/hadoop/yarn/lib/ehcache-3.3.1.jar:/server/hadoop/share/hadoop/yarn/lib/java-util-1.9.0.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-registry-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-nodemanager-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-sharedcachemanager-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-common-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-common-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-web-proxy-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-timeline-pluginstorage-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-tests-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-client-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-applications-mawo-core-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-router-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-services-api-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-api-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-services-core-3.3.6.jar:/server/hadoop/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-3.3.6.jar STARTUP_MSG: build = https://github.com/apache/hadoop.git -r 1be78238728da9266a4f88195058f08fd012bf9c; compiled by 'ubuntu' on 2023-06-18T08:22Z STARTUP_MSG: java = 1.8.0_401 ************************************************************/ 2025-08-23 21:29:44,272 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: registered UNIX signal handlers for [TERM, HUP, INT] 2025-08-23 21:29:44,656 INFO org.apache.hadoop.hdfs.server.datanode.checker.ThrottledAsyncChecker: Scheduling a check for [DISK]file:/server/hadoop/data/dn 2025-08-23 21:29:44,792 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: Loaded properties from hadoop-metrics2.properties 2025-08-23 21:29:44,866 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled Metric snapshot period at 10 second(s). 2025-08-23 21:29:44,866 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system started 2025-08-23 21:29:45,210 INFO org.apache.hadoop.hdfs.server.common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2025-08-23 21:29:45,231 INFO org.apache.hadoop.hdfs.server.datanode.BlockScanner: Initialized block scanner with targetBytesPerSec 1048576 2025-08-23 21:29:45,236 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Configured hostname is master 2025-08-23 21:29:45,236 INFO org.apache.hadoop.hdfs.server.common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2025-08-23 21:29:45,241 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Starting DataNode with maxLockedMemory = 0 2025-08-23 21:29:45,264 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened streaming server at /0.0.0.0:9866 2025-08-23 21:29:45,266 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwidth is 104857600 bytes/s 2025-08-23 21:29:45,266 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Number threads for balancing is 100 2025-08-23 21:29:45,304 INFO org.eclipse.jetty.util.log: Logging initialized @1787ms to org.eclipse.jetty.util.log.Slf4jLog 2025-08-23 21:29:45,503 WARN org.apache.hadoop.security.authentication.server.AuthenticationFilter: Unable to initialize FileSignerSecretProvider, falling back to use random secrets. Reason: Could not read signature secret file: /root/hadoop-http-auth-signature-secret 2025-08-23 21:29:45,513 INFO org.apache.hadoop.http.HttpRequestLog: Http request log for http.requests.datanode is not defined 2025-08-23 21:29:45,524 INFO org.apache.hadoop.http.HttpServer2: Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) 2025-08-23 21:29:45,527 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context datanode 2025-08-23 21:29:45,527 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2025-08-23 21:29:45,528 INFO org.apache.hadoop.http.HttpServer2: Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2025-08-23 21:29:45,572 INFO org.apache.hadoop.http.HttpServer2: Jetty bound to port 41547 2025-08-23 21:29:45,573 INFO org.eclipse.jetty.server.Server: jetty-9.4.51.v20230217; built: 2023-02-17T08:19:37.309Z; git: b45c405e4544384de066f814ed42ae3dceacdd49; jvm 1.8.0_401-b10 2025-08-23 21:29:45,614 INFO org.eclipse.jetty.server.session: DefaultSessionIdManager workerName=node0 2025-08-23 21:29:45,614 INFO org.eclipse.jetty.server.session: No SessionScavenger set, using defaults 2025-08-23 21:29:45,615 INFO org.eclipse.jetty.server.session: node0 Scavenging every 600000ms 2025-08-23 21:29:45,632 INFO org.eclipse.jetty.server.handler.ContextHandler: Started o.e.j.s.ServletContextHandler@4fad9bb2{logs,/logs,file:///server/hadoop/logs/,AVAILABLE} 2025-08-23 21:29:45,633 INFO org.eclipse.jetty.server.handler.ContextHandler: Started o.e.j.s.ServletContextHandler@1941a8ff{static,/static,file:///server/hadoop/share/hadoop/hdfs/webapps/static/,AVAILABLE} 2025-08-23 21:29:45,708 INFO org.eclipse.jetty.server.handler.ContextHandler: Started o.e.j.w.WebAppContext@35390ee3{datanode,/,file:///server/hadoop/share/hadoop/hdfs/webapps/datanode/,AVAILABLE}{file:/server/hadoop/share/hadoop/hdfs/webapps/datanode} 2025-08-23 21:29:45,720 INFO org.eclipse.jetty.server.AbstractConnector: Started ServerConnector@66565121{HTTP/1.1, (http/1.1)}{localhost:41547} 2025-08-23 21:29:45,720 INFO org.eclipse.jetty.server.Server: Started @2203ms 2025-08-23 21:29:45,796 WARN org.apache.hadoop.hdfs.server.datanode.web.DatanodeHttpServer: Got null for restCsrfPreventionFilter - will not do any filtering. 2025-08-23 21:29:45,892 INFO org.apache.hadoop.hdfs.server.datanode.web.DatanodeHttpServer: Listening HTTP traffic on /0.0.0.0:9864 2025-08-23 21:29:45,904 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: dnUserName = root 2025-08-23 21:29:45,905 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: supergroup = supergroup 2025-08-23 21:29:45,914 INFO org.apache.hadoop.util.JvmPauseMonitor: Starting JVM pause monitor 2025-08-23 21:29:46,034 INFO org.apache.hadoop.ipc.CallQueueManager: Using callQueue: class java.util.concurrent.LinkedBlockingQueue, queueCapacity: 1000, scheduler: class org.apache.hadoop.ipc.DefaultRpcScheduler, ipcBackoff: false. 2025-08-23 21:29:46,052 INFO org.apache.hadoop.ipc.Server: Listener at 0.0.0.0:9867 2025-08-23 21:29:46,055 INFO org.apache.hadoop.ipc.Server: Starting Socket Reader #1 for port 9867 2025-08-23 21:29:46,463 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Opened IPC server at /0.0.0.0:9867 2025-08-23 21:29:46,501 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Refresh request received for nameservices: null 2025-08-23 21:29:46,517 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: Unable to get NameNode addresses. java.io.IOException: Incorrect configuration: namenode address dfs.namenode.servicerpc-address.[] or dfs.namenode.rpc-address.[] is not configured. at org.apache.hadoop.hdfs.DFSUtil.getNNServiceRpcAddressesForCluster(DFSUtil.java:656) at org.apache.hadoop.hdfs.server.datanode.BlockPoolManager.refreshNamenodes(BlockPoolManager.java:157) at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1755) at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:564) at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:3148) at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:3054) at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:3098) at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:3242) at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:3266) 2025-08-23 21:29:46,558 INFO org.eclipse.jetty.server.handler.ContextHandler: Stopped o.e.j.w.WebAppContext@35390ee3{datanode,/,null,STOPPED}{file:/server/hadoop/share/hadoop/hdfs/webapps/datanode} 2025-08-23 21:29:46,565 INFO org.eclipse.jetty.server.AbstractConnector: Stopped ServerConnector@66565121{HTTP/1.1, (http/1.1)}{localhost:0} 2025-08-23 21:29:46,565 INFO org.eclipse.jetty.server.session: node0 Stopped scavenging 2025-08-23 21:29:46,565 INFO org.eclipse.jetty.server.handler.ContextHandler: Stopped o.e.j.s.ServletContextHandler@1941a8ff{static,/static,file:///server/hadoop/share/hadoop/hdfs/webapps/static/,STOPPED} 2025-08-23 21:29:46,565 INFO org.eclipse.jetty.server.handler.ContextHandler: Stopped o.e.j.s.ServletContextHandler@4fad9bb2{logs,/logs,file:///server/hadoop/logs/,STOPPED} 2025-08-23 21:29:46,573 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Waiting up to 30 seconds for transfer threads to complete 2025-08-23 21:29:46,574 INFO org.apache.hadoop.ipc.Server: Stopping server on 9867 2025-08-23 21:29:46,590 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping DataNode metrics system... 2025-08-23 21:29:46,590 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system stopped. 2025-08-23 21:29:46,590 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system shutdown complete. 2025-08-23 21:29:46,596 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: Shutdown complete. 2025-08-23 21:29:46,596 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: Exception in secureMain java.io.IOException: No services to connect, missing NameNode address. at org.apache.hadoop.hdfs.server.datanode.BlockPoolManager.refreshNamenodes(BlockPoolManager.java:165) at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1755) at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:564) at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:3148) at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:3054) at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:3098) at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:3242) at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:3266) 2025-08-23 21:29:46,598 INFO org.apache.hadoop.util.ExitUtil: Exiting with status 1: java.io.IOException: No services to connect, missing NameNode address. 2025-08-23 21:29:46,634 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down DataNode at master/192.168.88.8 ************************************************************/
最新发布
08-24
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值