HTTP Status 404 - No result defined for action com.action.personAction and result input

本文详细解析了Struts2框架中校验机制的工作原理,并通过实例演示了如何正确配置校验流程及解决常见的404错误。文章强调了在配置文件中正确指定input返回视图的重要性。

struts2中进行有效性校验时练习的小例子中产生下面的报错信息:
   页面显示:HTTP Status 404 - No result defined for action com.action.personAction and result input

  控制台显示:
     警告: Could not find action or result
     No result defined for action com.action.personAction and result input

出错原因是:
  1)前台请求页面的属性值和action中对应的属性值写错了个单词,导致action获取不到,属性值为空。
Action中的属性值为空,Struts2的默认拦截器会报错,找不到input的result,不能把错误返回。只能报找不到result.

 
  2)在进行校验时,在配置文件中没有指定input返回视图

       系统包含fieldErrors失败信息时,struts2会将请求转发到name为input的result。在input视图中可以显示校验不通过的报错提示信息。通过:<%@taglib uri="/struts-tags" prefix="s"%>       <s:fielderror/>
         <action name = "validate_*" class = "com.action.personAction" method = "save">
              <result name="input">/index.jsp</result>  <!--少这行报错 -->
              <result name = "success" >/WEB-INF/page/validate.jsp</result>
         </action>

导致这个报错的原因太多...



21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.federation.gpg.application.cleaner.interval-ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.output.fileoutputformat.compress 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.federation.gpg.subcluster.cleaner.interval-ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.sharedcache.store.in-memory.staleness-period-mins 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client.write.byte-array-manager.count-limit 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.runtime.linux.runc.layer-mounts-to-keep 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.group.mapping.providers.combined 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.job.running.map.limit 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.webapp.address 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.placement-constraints.scheduler.pool-size 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.s3a.multipart.size 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client.slow.io.warning.threshold.ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.app.mapreduce.am.job.committer.commit-window 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.job.submithostname 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.edits.asynclogging 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.blockreport.incremental.intervalMsec 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.ifile.readahead 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.federation.state-store.sql.conn-time-out 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.runtime.linux.runc.image-tag-to-manifest-plugin 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client.socketcache.capacity 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.s3a.select.input.csv.field.delimiter 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client.retry.policy.spec 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.reencrypt.batch.size 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.s3a.connection.ssl.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.proxyuser.hadoop.hosts 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.read.considerLoad 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.datanode.max.slowdisks.to.exclude 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.groups.cache.secs 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.datanode.peer.stats.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.replication 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.storage.policy.satisfier.work.multiplier.per.iteration 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.group.mapping.ldap.directory.search.timeout 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.datanode.available-space-volume-choosing-policy.balanced-space-threshold 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.checksum.combine.mode 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.storage.policy.satisfier.max.outstanding.paths 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.sleep-delay-before-sigkill.ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.apps.cache.enable 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.automatic.close 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.reencrypt.edek.threads 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.disk-health-checker.disk-free-space-threshold.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.acls.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client.short.circuit.replica.stale.threshold.ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.health-checker.run-before-startup 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.send.qop.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.jobhistory.intermediate-done-dir 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.slowpeer.collect.interval 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client.server-defaults.validity.period.ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.client.libjars.wildcard 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.storage.policy.satisfier.address 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.reduce.shuffle.input.buffer.percent 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.audit.loggers 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for io.serializations 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.dispatcher.print-thread-pool.keep-alive-time 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.http.cross-origin.allowed-methods 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.snapshot.capture.openfiles 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.qjournal.queued-edits.limit.mb 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.zk.acl 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.container.stderr.pattern 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.cluster.local.dir 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for ipc.[port_number].cost-provider.impl 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.kerberos.kinit.command 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.metrics.logger.period.seconds 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.viewfs.overload.scheme.target.abfss.impl 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.block.access.token.lifetime 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.delegation.token.max-lifetime 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.datanode.drop.cache.behind.writes 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.system-metrics-publisher.timeline-server-v1.enable-batch 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.remove.dead.datanode.batchnum 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.submission-preprocessor.file-refresh-interval-ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.num.extra.edits.retained 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.block.placement.ec.classname 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for ipc.client.connect.max.retries.on.timeouts 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.client.resolve.topology.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.qjournal.http.open.timeout.ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for ha.health-monitor.connect-retry-interval.ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.edekcacheloader.initial.delay.ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client.rbf.observer.read.enable 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client.failover.resolver.useFQDN 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for io.mapfile.bloom.size 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.ftp.data.connection.mode 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client-write-packet-size 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.app.mapreduce.shuffle.log.backups 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.kerberos.principal.pattern 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.webhdfs.socket.connect-timeout 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.scheduler.monitor.enable 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.proxyuser.hadoop.groups 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.s3a.select.output.csv.quote.character 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.task.stuck.timeout-ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.authorization 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.timeline-service.version 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.am.liveness-monitor.expiry-interval-ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.federation.gpg.webapp.address 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.timeline-service.leveldb-timeline-store.path 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.job.reduce.slowstart.completedmaps 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.delegation.token.max-lifetime 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.ha.automatic-failover.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.datanode.socket.write.timeout 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.accesstime.precision 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.group.mapping.ldap.conversion.rule 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for io.mapfile.bloom.error.rate 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.webapp.rest-csrf.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.timeline-service.leveldb-state-store.path 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.scheduler.configuration.zk-store.parent-path 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for ipc.[port_number].backoff.enable 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.timeline-service.writer.flush-interval-seconds 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.posix.acl.inheritance.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.datanode.outliers.report.interval 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.kms.client.encrypted.key.cache.low-watermark 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.top.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.s3a.retry.throttle.interval 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.jobhistory.webapp.rest-csrf.custom-header 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.webapp.xfs-filter.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for ipc.identity-provider.impl 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client.cached.conn.retry 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.submission-preprocessor.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.system.tags 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.runtime.linux.runc.image-tag-to-manifest-plugin.num-manifests-to-cache 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.least-load-policy-selector.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.numa-awareness.numactl.cmd 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.path.based.cache.refresh.interval.ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.fs-limits.max-directory-items 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.ha.log-roll.period 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.distributed-scheduling.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.datanode.pmem.cache.recovery 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.minicluster.fixed.ports 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.storage.policy.satisfier.queue.limit 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.snapshot.filesystem.limit 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.resource.percentage-physical-cpu-limit 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.fs-limits.max-xattr-size 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.blocks.per.postponedblocks.rescan 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.maintenance.replication.min 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.timeline-service.app-aggregation-interval-secs 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.max.op.size 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.iostatistics.thread.level.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.job.reducer.unconditional-preempt.delay.sec 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.app.mapreduce.am.hard-kill-timeout-ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.s3a.connection.ttl 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.storage.policy.permissions.superuser-only 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.df.interval 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.job.cache.limit.max-single-resource-mb 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.s3a.assumed.role.session.duration 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.disk.balancer.block.tolerance.percent 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.webhdfs.netty.high.watermark 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.datanode.balance.max.concurrent.moves 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.log.delete.threshold 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.job.token.tracking.ids.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.s3a.assumed.role.credentials.provider 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.log-container-debug-info-on-error.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.kms.client.failover.sleep.max.millis 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.webapp.rest-csrf.custom-header 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.jobhistory.move.thread-count 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for io.compression.codec.zstd.level 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.timeline-service.http-authentication.simple.anonymous.allowed 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.provided.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.sharedcache.client-server.thread-count 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.scheduler.configuration.max.version 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.jobhistory.jobname.limit 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.dispatcher.print-events-info.threshold 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.decommission.blocks.per.interval 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.qjournal.write-txns.timeout.ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.federation.subcluster-resolver.class 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.read-lock-reporting-threshold-ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.task.timeout 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.resource.memory-mb 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.container-log-monitor.total-size-limit-bytes 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.fileoutputcommitter.algorithm.version 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.framework.name 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.router.clientrm.interceptor-class.pipeline 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.system-metrics-publisher.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.sharedcache.nested-level 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.dns.log-slow-lookups.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.jobhistory.webapp.https.address 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for file.client-write-packet-size 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for ipc.client.ping 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.federation.state-store.sql.idle-time-out 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.federation.gpg.policy.generator.interval 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.federation.gpg.webapp.https.address 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.balancer.max-no-move-interval 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.minicluster.control-resource-monitoring 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.disk.balancer.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.fs.state-store.num-retries 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.uid.cache.secs 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.ha.automatic-failover.zk-base-path 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.job.speculative.speculative-cap-running-tasks 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.node-labels.am.allow-non-exclusive-allocation 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.datanode.du.reserved.calculator 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.datanode.block.id.layout.upgrade.threads 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for io.erasurecode.codec.native.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.client.load.resource-types.from-server 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.client.application-client-protocol.poll-timeout-ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.datanode.oob.timeout-ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.job.sharedcache.mode 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.job.hdfs-servers 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.epoch.range 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.federation.gpg.subcluster.heartbeat.expiration-ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.map.output.compress 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.token.service.use_ip 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.kms.client.encrypted.key.cache.num.refill.threads 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.edekcacheloader.interval.ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.aux-services.mapreduce_shuffle.class 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.group.mapping.ldap.num.attempts.before.failover 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.du.interval 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client.read.uri.cache.enabled 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.zk.retry-interval-ms 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.data.transfer.server.tcpnodelay 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.job.dir 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.http.client.failover.max.attempts 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.s3a.socket.send.buffer 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.client.block.write.locateFollowingBlock.retries 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.jvm.system-properties-to-log 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.enable.retrycache 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.job.encrypted-intermediate-data.buffer.kb 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.nodemanager.resource-plugins.gpu.docker-plugin.nvidia-docker-v1.endpoint 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.data.transfer.client.tcpnodelay 21:37:31.523 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.storage.policy.satisfier.mode 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.webapp.xfs-filter.xframe-options 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.reduce.memory.mb 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.caller.context.enabled 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.nodemanagers.heartbeat-interval-speedup-factor 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.qjournal.prepare-recovery.timeout.ms 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.router.deregister.subcluster.enabled 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.security.sensitive-config-keys 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for mapreduce.client.completion.pollinterval 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.namenode.secondary.http-address 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.router.interceptor.allow-partial-result.enable 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for yarn.resourcemanager.webapp.https.address 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for fs.s3a.retry.throttle.limit 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for dfs.permissions.allow.owner.set.quota 21:37:31.524 [main] DEBUG org.apache.hadoop.conf.Configuration - Handling deprecation for hadoop.domainname.resolver.impl 21:37:31.597 [main] INFO org.apache.hadoop.mapreduce.Job - The url to track the job: http://localhost:8080/ 21:37:31.598 [main] INFO org.apache.hadoop.mapreduce.Job - Running job: job_local1106899704_0001 21:37:31.601 [Thread-5] INFO org.apache.hadoop.mapred.LocalJobRunner - OutputCommitter set in config null 21:37:31.603 [main] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.mapreduce.Job$1@7c6442c2] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.mapreduce.Job.updateStatus(Job.java:329) at org.apache.hadoop.mapreduce.Job.isComplete(Job.java:613) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1736) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1698) at cn.itcast.mr.dedup.MatrixMultiplication.main(MatrixMultiplication.java:128) 21:37:31.611 [main] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.mapreduce.Job$1@2098d37d] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.mapreduce.Job.updateStatus(Job.java:329) at org.apache.hadoop.mapreduce.Job.isComplete(Job.java:613) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1737) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1698) at cn.itcast.mr.dedup.MatrixMultiplication.main(MatrixMultiplication.java:128) 21:37:31.612 [Thread-5] DEBUG org.apache.hadoop.mapreduce.lib.output.PathOutputCommitterFactory - Looking for committer factory for path hdfs://192.168.88.101:8020/output 21:37:31.612 [Thread-5] DEBUG org.apache.hadoop.mapreduce.lib.output.PathOutputCommitterFactory - No scheme-specific factory defined in mapreduce.outputcommitter.factory.scheme.hdfs 21:37:31.612 [Thread-5] INFO org.apache.hadoop.mapreduce.lib.output.PathOutputCommitterFactory - No output committer factory defined, defaulting to FileOutputCommitterFactory 21:37:31.613 [Thread-5] DEBUG org.apache.hadoop.mapreduce.lib.output.PathOutputCommitterFactory - Creating FileOutputCommitter for path hdfs://192.168.88.101:8020/output and context TaskAttemptContextImpl{JobContextImpl{jobId=job_local1106899704_0001}; taskId=attempt_local1106899704_0001_m_000000_0, status=''} 21:37:31.613 [Thread-5] DEBUG org.apache.hadoop.mapreduce.lib.output.PathOutputCommitter - Instantiating committer FileOutputCommitter{PathOutputCommitter{context=TaskAttemptContextImpl{JobContextImpl{jobId=job_local1106899704_0001}; taskId=attempt_local1106899704_0001_m_000000_0, status=''}; org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter@9f9d7ce}; outputPath=null, workPath=null, algorithmVersion=0, skipCleanup=false, ignoreCleanupFailures=false} with output path hdfs://192.168.88.101:8020/output and job context TaskAttemptContextImpl{JobContextImpl{jobId=job_local1106899704_0001}; taskId=attempt_local1106899704_0001_m_000000_0, status=''} 21:37:31.614 [Thread-5] INFO org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter - File Output Committer Algorithm version is 2 21:37:31.614 [Thread-5] INFO org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter - FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false 21:37:31.615 [Thread-5] INFO org.apache.hadoop.mapred.LocalJobRunner - OutputCommitter is org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter 21:37:31.622 [Thread-5] DEBUG org.apache.hadoop.fs.statistics.impl.IOStatisticsContextIntegration - Created instance IOStatisticsContextImpl{id=2, threadId=32, ioStatistics=counters=(); gauges=(); minimums=(); maximums=(); means=(); } 21:37:31.629 [Thread-5] DEBUG org.apache.hadoop.hdfs.DFSClient - /output/_temporary/0: masked={ masked: rwxr-xr-x, unmasked: rwxrwxrwx } 21:37:31.637 [IPC Parameter Sending Thread for xxjdxnj/192.168.88.101:8020] DEBUG org.apache.hadoop.ipc.Client - IPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from СIPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from С sending #3 org.apache.hadoop.hdfs.protocol.ClientProtocol.mkdirs 21:37:31.649 [IPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from С] DEBUG org.apache.hadoop.ipc.Client - IPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from С got value #3 21:37:31.654 [Thread-5] DEBUG org.apache.hadoop.io.retry.RetryInvocationHandler - Exception while invoking call #3 ClientNamenodeProtocolTranslatorPB.mkdirs over null. Not retrying because try once and fail. org.apache.hadoop.ipc.RemoteException: Permission denied: user=С, access=WRITE, inode="/":hadoop:supergroup:drwxr-xr-x at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:661) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:501) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermissionWithContext(FSPermissionChecker.java:525) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:395) at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1964) at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1945) at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAncestorAccess(FSDirectory.java:1904) at org.apache.hadoop.hdfs.server.namenode.FSDirMkdirOp.mkdirs(FSDirMkdirOp.java:60) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3531) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:1173) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:750) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3203) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) at org.apache.hadoop.ipc.Client.call(Client.java:1529) at org.apache.hadoop.ipc.Client.call(Client.java:1426) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) at jdk.proxy2/jdk.proxy2.$Proxy11.mkdirs(Unknown Source) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$mkdirs$20(ClientNamenodeProtocolTranslatorPB.java:611) at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:611) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) at jdk.proxy2/jdk.proxy2.$Proxy12.mkdirs(Unknown Source) at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2555) at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2531) at org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1497) at org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1494) at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:1511) at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:1486) at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:2494) at org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.setupJob(FileOutputCommitter.java:356) at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:541) 21:37:31.663 [IPC Parameter Sending Thread for xxjdxnj/192.168.88.101:8020] DEBUG org.apache.hadoop.ipc.Client - IPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from СIPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from С sending #4 org.apache.hadoop.hdfs.protocol.ClientProtocol.delete 21:37:31.674 [IPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from С] DEBUG org.apache.hadoop.ipc.Client - IPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from С got value #4 21:37:31.675 [Thread-5] DEBUG org.apache.hadoop.ipc.ProtobufRpcEngine2 - Call: delete took 12ms 21:37:31.678 [Thread-5] WARN org.apache.hadoop.mapred.LocalJobRunner - job_local1106899704_0001 org.apache.hadoop.security.AccessControlException: Permission denied: user=С, access=WRITE, inode="/":hadoop:supergroup:drwxr-xr-x at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:661) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:501) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermissionWithContext(FSPermissionChecker.java:525) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:395) at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1964) at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1945) at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAncestorAccess(FSDirectory.java:1904) at org.apache.hadoop.hdfs.server.namenode.FSDirMkdirOp.mkdirs(FSDirMkdirOp.java:60) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3531) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:1173) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:750) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3203) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.base/java.lang.reflect.Constructor.newInstanceWithCaller(Constructor.java:499) at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:480) at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:121) at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:88) at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2557) at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2531) at org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1497) at org.apache.hadoop.hdfs.DistributedFileSystem$27.doCall(DistributedFileSystem.java:1494) at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:1511) at org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:1486) at org.apache.hadoop.fs.FileSystem.mkdirs(FileSystem.java:2494) at org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.setupJob(FileOutputCommitter.java:356) at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:541) Caused by: org.apache.hadoop.ipc.RemoteException: Permission denied: user=С, access=WRITE, inode="/":hadoop:supergroup:drwxr-xr-x at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.check(FSPermissionChecker.java:661) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:501) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermissionWithContext(FSPermissionChecker.java:525) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:395) at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1964) at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkPermission(FSDirectory.java:1945) at org.apache.hadoop.hdfs.server.namenode.FSDirectory.checkAncestorAccess(FSDirectory.java:1904) at org.apache.hadoop.hdfs.server.namenode.FSDirMkdirOp.mkdirs(FSDirMkdirOp.java:60) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3531) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:1173) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:750) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:621) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:589) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine2.java:573) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1227) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1246) at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:1169) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1953) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:3203) at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1584) at org.apache.hadoop.ipc.Client.call(Client.java:1529) at org.apache.hadoop.ipc.Client.call(Client.java:1426) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:258) at org.apache.hadoop.ipc.ProtobufRpcEngine2$Invoker.invoke(ProtobufRpcEngine2.java:139) at jdk.proxy2/jdk.proxy2.$Proxy11.mkdirs(Unknown Source) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.lambda$mkdirs$20(ClientNamenodeProtocolTranslatorPB.java:611) at org.apache.hadoop.ipc.internal.ShadedProtobufHelper.ipc(ShadedProtobufHelper.java:160) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.mkdirs(ClientNamenodeProtocolTranslatorPB.java:611) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:437) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:170) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:162) at org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:100) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:366) at jdk.proxy2/jdk.proxy2.$Proxy12.mkdirs(Unknown Source) at org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:2555) ... 9 common frames omitted 21:37:31.683 [Thread-5] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.fs.FileContext$2@15fc336f] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:343) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:465) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:442) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:428) at org.apache.hadoop.mapred.LocalDistributedCacheManager.close(LocalDistributedCacheManager.java:268) at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:598) 21:37:32.626 [main] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.mapreduce.Job$1@77b9d0c7] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.mapreduce.Job.updateStatus(Job.java:329) at org.apache.hadoop.mapreduce.Job.isUber(Job.java:1866) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1747) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1698) at cn.itcast.mr.dedup.MatrixMultiplication.main(MatrixMultiplication.java:128) 21:37:32.626 [main] INFO org.apache.hadoop.mapreduce.Job - Job job_local1106899704_0001 running in uber mode : false 21:37:32.628 [main] INFO org.apache.hadoop.mapreduce.Job - map 0% reduce 0% 21:37:32.628 [main] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.mapreduce.Job$6@3b0ee03a] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.mapreduce.Job.getTaskCompletionEvents(Job.java:730) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1759) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1698) at cn.itcast.mr.dedup.MatrixMultiplication.main(MatrixMultiplication.java:128) 21:37:32.629 [main] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.mapreduce.Job$1@796065aa] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.mapreduce.Job.updateStatus(Job.java:329) at org.apache.hadoop.mapreduce.Job.isComplete(Job.java:613) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1736) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1698) at cn.itcast.mr.dedup.MatrixMultiplication.main(MatrixMultiplication.java:128) 21:37:32.629 [main] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.mapreduce.Job$1@28a6301f] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.mapreduce.Job.updateStatus(Job.java:329) at org.apache.hadoop.mapreduce.Job.isComplete(Job.java:613) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1737) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1698) at cn.itcast.mr.dedup.MatrixMultiplication.main(MatrixMultiplication.java:128) 21:37:32.630 [main] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.mapreduce.Job$6@2c306a57] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.mapreduce.Job.getTaskCompletionEvents(Job.java:730) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1759) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1698) at cn.itcast.mr.dedup.MatrixMultiplication.main(MatrixMultiplication.java:128) 21:37:32.630 [main] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.mapreduce.Job$1@773e2eb5] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.mapreduce.Job.updateStatus(Job.java:329) at org.apache.hadoop.mapreduce.Job.isComplete(Job.java:613) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1736) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1698) at cn.itcast.mr.dedup.MatrixMultiplication.main(MatrixMultiplication.java:128) 21:37:32.631 [main] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.mapreduce.Job$1@d8948cd] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.mapreduce.Job.updateStatus(Job.java:329) at org.apache.hadoop.mapreduce.Job.isSuccessful(Job.java:625) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1763) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1698) at cn.itcast.mr.dedup.MatrixMultiplication.main(MatrixMultiplication.java:128) 21:37:32.631 [main] INFO org.apache.hadoop.mapreduce.Job - Job job_local1106899704_0001 failed with state FAILED due to: NA 21:37:32.631 [main] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.mapreduce.Job$8@7abe27bf] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:818) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1770) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1698) at cn.itcast.mr.dedup.MatrixMultiplication.main(MatrixMultiplication.java:128) 21:37:32.651 [main] INFO org.apache.hadoop.mapreduce.Job - Counters: 0 21:37:32.651 [main] DEBUG org.apache.hadoop.security.UserGroupInformation - PrivilegedAction [as: С (auth:SIMPLE)][action: org.apache.hadoop.mapreduce.Job$1@2679311f] java.lang.Exception: null at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1950) at org.apache.hadoop.mapreduce.Job.updateStatus(Job.java:329) at org.apache.hadoop.mapreduce.Job.isSuccessful(Job.java:625) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1710) at cn.itcast.mr.dedup.MatrixMultiplication.main(MatrixMultiplication.java:128) 21:37:32.653 [shutdown-hook-0] DEBUG org.apache.hadoop.fs.FileSystem - FileSystem.close() by method: org.apache.hadoop.hdfs.DistributedFileSystem.close(DistributedFileSystem.java:1530)); Key: (С (auth:SIMPLE))@hdfs://192.168.88.101:8020; URI: hdfs://192.168.88.101:8020; Object Identity Hash: 2e075efe 21:37:32.653 [shutdown-hook-0] DEBUG org.apache.hadoop.ipc.Client - stopping client from cache: Client-e9ac678cebb441d58dd3dc3f8f54b798 21:37:32.654 [shutdown-hook-0] DEBUG org.apache.hadoop.ipc.Client - removing client from cache: Client-e9ac678cebb441d58dd3dc3f8f54b798 21:37:32.654 [shutdown-hook-0] DEBUG org.apache.hadoop.ipc.Client - stopping actual client because no more references remain: Client-e9ac678cebb441d58dd3dc3f8f54b798 21:37:32.654 [shutdown-hook-0] DEBUG org.apache.hadoop.ipc.Client - Stopping client 21:37:32.655 [IPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from С] DEBUG org.apache.hadoop.ipc.Client - IPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from С: closed 21:37:32.655 [IPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from С] DEBUG org.apache.hadoop.ipc.Client - IPC Client (1759899303) connection to xxjdxnj/192.168.88.101:8020 from С: stopped, remaining connections 0 21:37:32.655 [shutdown-hook-0] DEBUG org.apache.hadoop.fs.FileSystem - FileSystem.close() by method: org.apache.hadoop.fs.FilterFileSystem.close(FilterFileSystem.java:529)); Key: (С (auth:SIMPLE))@file://; URI: file:///; Object Identity Hash: 2a38dfe6 21:37:32.655 [shutdown-hook-0] DEBUG org.apache.hadoop.fs.FileSystem - FileSystem.close() by method: org.apache.hadoop.fs.RawLocalFileSystem.close(RawLocalFileSystem.java:895)); Key: null; URI: file:///; Object Identity Hash: 6f3a54c5 21:37:32.656 [shutdown-hook-0] DEBUG org.apache.hadoop.hdfs.KeyProviderCache - Invalidating all cached KeyProviders. 21:37:32.656 [Thread-1] DEBUG org.apache.hadoop.util.ShutdownHookManager - Completed shutdown in 0.004 seconds; Timeouts: 0 21:37:32.664 [Thread-1] DEBUG org.apache.hadoop.util.ShutdownHookManager - ShutdownHookManager completed shutdown. Process finished with exit code 1
06-22
element-ui.common.js:29286 POST http://localhost:9876/uplodPython 404 (Not Found) upload @ element-ui.common.js:29286 post @ element-ui.common.js:29592 upload @ element-ui.common.js:29520 eval @ element-ui.common.js:29511 uploadFiles @ element-ui.common.js:29509 handleChange @ element-ui.common.js:29490 invokeWithErrorHandling @ vue.runtime.esm.js:3058 invoker @ vue.runtime.esm.js:1859 original_1._wrapper @ vue.runtime.esm.js:7508 index.js??clonedRuleSet-40.use[0]!./node_modules/@vue/cli-service/node_modules/@vue/vue-loader-v15/lib/index.js??vue-loader-options!./src/views/uplodPython.vue?vue&type=script&lang=js:105 Uncaught (in promise) ReferenceError: res is not defined at eval (index.js??clonedRuleSet-40.use[0]!./node_modules/@vue/cli-service/node_modules/@vue/vue-loader-v15/lib/index.js??vue-loader-options!./src/views/uplodPython.vue?vue&type=script&lang=js:105:21) eval @ index.js??clonedRuleSet-40.use[0]!./node_modules/@vue/cli-service/node_modules/@vue/vue-loader-v15/lib/index.js??vue-loader-options!./src/views/uplodPython.vue?vue&type=script&lang=js:105 Promise.then handleFileUpload @ index.js??clonedRuleSet-40.use[0]!./node_modules/@vue/cli-service/node_modules/@vue/vue-loader-v15/lib/index.js??vue-loader-options!./src/views/uplodPython.vue?vue&type=script&lang=js:101 handleError @ element-ui.common.js:29887 onError @ element-ui.common.js:29588 onload @ element-ui.common.js:29267 XMLHttpRequest.send upload @ element-ui.common.js:29286 post @ element-ui.common.js:29592 upload @ element-ui.common.js:29520 eval @ element-ui.common.js:29511 uploadFiles @ element-ui.common.js:29509 handleChange @ element-ui.common.js:29490 invokeWithErrorHandling @ vue.runtime.esm.js:3058 invoker @ vue.runtime.esm.js:1859 original_1._wrapper @ vue.runtime.esm.js:7508 index.js??clonedRuleSet-40.use[0]!./node_modules/@vue/cli-service/node_modules/@vue/vue-loader-v15/lib/index.js??vue-loader-options!./src/views/uplodPython.vue?vue&type=script&lang=js:105 Uncaught (in promise) ReferenceError: res is not defined at eval (index.js??clonedRuleSet-40.use[0]!./node_modules/@vue/cli-service/node_modules/@vue/vue-loader-v15/lib/index.js??vue-loader-options!./src/views/uplodPython.vue?vue&type=script&lang=js:105:21)<template> <div class="classifier-container"> <el-card class="box-card"> <h2 style="text-align: center; color: #1890ff">🐾 动物图像分类器</h2> <p style="text-align: center; color: #666; margin-bottom: 30px"> 支持识别:猫(cat)、狗(dog)、鱼(fish)、大象(elephant) </p> <!-- 图片上传区 --> <el-upload class="upload-box" drag action="#" :auto-upload="true" :show-file-list="false" :on-change="handleFileUpload" accept="image/jpeg, image/jpg, image/png" > <i class="el-icon-upload"></i> <div class="el-upload__text"> 将图片拖到此处,或点击选择图片 </div> <div class="el-upload__tip" slot="tip"> 支持 JPG、PNG 格式,大小不超过 5MB </div> </el-upload> <!-- 预览与加载状态 --> <div v-if="previewImage" class="preview-section"> <h4>🖼 图像预览</h4> <img :src="previewImage" alt="预览" class="preview-img" /> <div v-if="loading" class="loading-box"> <el-progress type="circle" :percentage="progress" :width="80" ></el-progress> <p style="margin-top: 10px; color: #409eff">推理中...</p> </div> <!-- 预测结果展示 --> <div v-if="result && !loading" class="result-section"> <h4>✅ 识别结果</h4> <el-alert :title="`检测到: ${result.predicted_class} (置信度: ${result.confidence})`" :type="getClassType(result.predicted_class)" show-icon :closable="false" ></el-alert> <!-- 概率分布图表 --> <div class="chart-box"> <h5>📊 各类别概率分布:</h5> <el-bar-chart :data="chartData" /> </div> </div> </div> </el-card> </div> </template> <script> // 简易条形图组件(无需 echarts,使用纯 CSS + Element UI) const ElBarChart = { props: ["data"], render(h) { return h( "div", { style: "margin-top: 10px;" }, this.data.map((item) => h("div", { style: "margin: 12px 0;" }, [ h( "span", { style: "display:inline-block;width:80px;color:#666;" }, item.label ), h("el-progress", { props: { percentage: item.value * 100, strokeWidth: 14, "text-inside": true, status: item.isMax ? "success" : "", }, style: "width: calc(100% - 90px); display:inline-block; vertical-align: middle;", }), ]) ) ); }, }; export default { components: { ElBarChart }, data() { return { previewImage: null, // 预览图片 base64 loading: false, // 是否正在推理 progress: 0, // 进度条 result: null, // 推理返回结果 uploadUrl: "http://localhost:9876/api/uploadPython", // Spring Boot 后端地址 classColors: { cat: "success", dog: "primary", fish: "info", elephant: "warning", }, }; }, computed: { chartData() { if (!this.result || !this.result.all_probabilities) return []; const all = this.result.all_probabilities; const maxClass = this.result.predicted_class; return Object.keys(all).map((label) => ({ label, value: all[label], isMax: label === maxClass, })); }, }, methods: { getClassType(className) { return this.classColors[className] || "info"; }, handleFileUpload(file, fileList) { const allowedTypes = ["image/jpeg", "image/jpg", "image/png"]; const maxSize = 5 * 1024 * 1024; // 5MB if (!allowedTypes.includes(file.raw.type)) { this.$message.error("仅支持 JPG/PNG 格式!"); return; } if (file.size > maxSize) { this.$message.error("图片不能超过 5MB!"); return; } // 显示预览 const reader = new FileReader(); reader.onload = (e) => { this.previewImage = e.target.result; }; reader.readAsDataURL(file.raw); // 重置状态 this.loading = true; this.progress = 0; this.result = null; // 模拟进度增长 const interval = setInterval(() => { if (this.progress >= 90) clearInterval(interval); else this.progress += 5; }, 200); // 创建 FormData 发送图片 const formData = new FormData(); formData.append("file", file.raw); // 调用后端 API fetch(this.uploadUrl, { method: "POST", body: formData, }) .then((res) => res.json()) .then((data) => { clearInterval(interval); this.progress = 100; this.loading = false; console.log(res); if (data.success) { this.result = data; this.$message.success(`识别完成:${data.predicted_class}`); } else { this.handleError(data); } }); // .catch((err) => { // clearInterval(interval); // this.loading = false; // this.handleError({ error: res.error }); // }); }, handleError(data) { console.error("推理失败:", data); this.$alert(data.error || "未知错误", "❌ 推理失败", { confirmButtonText: "确定", type: "error", }); }, }, }; </script> <style scoped> .classifier-container { max-width: 800px; margin: 40px auto; padding: 20px; } .box-card { border-radius: 12px; box-shadow: 0 4px 12px rgba(0, 0, 0, 0.1); } .upload-box { width: 100%; margin-bottom: 30px; } .preview-section { margin-top: 20px; text-align: center; } .preview-img { max-width: 100%; max-height: 300px; border-radius: 8px; box-shadow: 0 2px 8px rgba(0, 0, 0, 0.1); margin-bottom: 20px; } .loading-box { margin: 20px auto; width: fit-content; } .result-section { text-align: left; margin-top: 20px; } .chart-box { margin-top: 20px; background: #f9f9f9; padding: 15px; border-radius: 8px; border: 1px solid #eee; } </style> package com.shop.jieyou.controller; import com.fasterxml.jackson.core.JsonProcessingException; import com.fasterxml.jackson.databind.JsonNode; import com.fasterxml.jackson.databind.ObjectMapper; import com.shop.jieyou.common.Result; import com.shop.jieyou.entity.UserItem; import com.shop.jieyou.service.ItemService; import com.shop.jieyou.service.PythonService; import org.springframework.beans.factory.annotation.Autowired; import org.springframework.http.ResponseEntity; import org.springframework.web.bind.annotation.*; import org.springframework.web.multipart.MultipartFile; import java.io.BufferedReader; import java.io.File; import java.io.IOException; import java.io.InputStreamReader; import java.util.ArrayList; import java.util.HashMap; import java.util.List; import java.util.Map; import java.util.concurrent.TimeUnit; /** * 花卉相关接口控制器 * 提供三大功能: * 1. 获取中国十大名花数据(来自爬虫或缓存) * 2. 手动刷新花卉数据(强制重新爬取) * 3. 基于用户行为的花卉推荐(调用Python协同过滤脚本) */ @RestController @CrossOrigin(origins = "*") // 允许所有域访问,用于前端开发调试(生产环境建议限制域名) @RequestMapping("/api") public class FlowerController { @Autowired private PythonService pythonService; // 注入业务服务层,处理数据获取与推荐逻辑 @Autowired private ItemService itemService; /** * 接口:GET /api/flowers * 功能:获取“中国十大名花”数据列表 * 数据来源:可能来自数据库、Redis 缓存 或 调用 Python 爬虫脚本 * * @return Result<List<Map<String, Object>>> 返回包含花卉信息的成功响应 */ @GetMapping("/flowers") public Result<List<Map<String, Object>>> getTopTenFlowers() { try { // 调用服务层获取花卉数据(内部可能带缓存机制) List<Map<String, Object>> flowers = pythonService.getFlowers(); return Result.success(flowers); // 成功返回数据 } catch (Exception e) { // 捕获异常并统一返回错误码和消息,避免暴露堆栈给前端 return Result.error("500", "获取花卉数据失败:" + e.getMessage()); } } /** * 接口:POST /api/flowers/refresh * 功能:强制刷新花卉数据缓存,触发重新爬取 * 使用场景:管理员手动更新数据时调用 * * @return Result<Map<String, Object>> 返回刷新结果信息 */ @PostMapping("/flowers/refresh") public Result<Map<String, Object>> refreshData() { try { // TODO: 如果实现了 clearCache 方法,请取消注释并调用 // pythonService.clearCache(); // 清除旧缓存,下次 getFlowers 将重新爬取 // 重新获取最新数据(假设此时会触发爬虫) List<Map<String, Object>> flowers = pythonService.getFlowers(); // 构造返回信息 Map<String, Object> data = new HashMap<>(); data.put("message", "数据已刷新"); data.put("count", flowers.size()); return Result.success(data); } catch (Exception e) { return Result.error("500", "刷新失败:" + e.getMessage()); } } // ========== 推荐系统相关常量定义 ========== /** * 输入文件路径:Java 将用户-商品行为数据写入此 JSON 文件供 Python 脚本读取 * 注意:src/main/resources 是编译后打包进 jar 的资源目录,不适合运行时写入! * 建议改为外部路径如 "./data/input.json" */ private static final String INPUT_PATH = "src/main/resources/scripts/input.json"; /** * 输出文件路径:Python 脚本将推荐结果写入此文件,Java 再读取返回给前端 */ private static final String OUTPUT_PATH = "src/main/resources/scripts/output.json"; /** * Python 协同过滤脚本路径 * 注意:resources 目录下的 .py 文件在打包后无法直接作为可执行脚本运行 * 更佳做法是将脚本放在项目外部或使用 ProcessBuilder 启动独立服务 */ private static final String PYTHON_SCRIPT = "src/main/resources/scripts/collaborative.py"; /** * 接口:GET /api/recommend?userId=123 * 功能:为指定用户生成个性化花卉推荐列表 * 实现方式:Java 查询数据库 → 写入 JSON 文件 → 调用 Python 脚本计算 → 读取结果返回 * * @param userId 用户ID,必填参数 * @return Result<JsonNode> 推荐的商品ID数组(如 [101, 105, 108]) */ @GetMapping("/recommend") public Result recommendFlowers(@RequestParam("userId") Long userId) { try { // 1. 获取用户行为数据 List<UserItem> matrix = pythonService.getUserItemMatrix(); // 2. 调用 Python 脚本(通过 stdin/stdout 通信) ProcessBuilder pb = new ProcessBuilder("python", PYTHON_SCRIPT, String.valueOf(userId)); pb.redirectErrorStream(true); // 合并错误流 Process process = pb.start(); // 3. 将数据写入脚本的标准输入 ObjectMapper mapper = new ObjectMapper(); mapper.writeValue(process.getOutputStream(), matrix); process.getOutputStream().close(); // 关闭输入,通知Python结束读取 // 4. 读取Python脚本的输出(推荐结果) JsonNode result = mapper.readTree(process.getInputStream()); // 5. 等待脚本执行完毕 int exitCode = process.waitFor(); if (exitCode != 0) { return Result.error("500", "Python script failed with exit code: " + exitCode); } System.out.println(result); return Result.success(result); } catch (Exception e) { e.printStackTrace(); return Result.error("500", "推荐生成失败:" + e.getMessage()); } } @PostMapping("/predict") public ResponseEntity<String> predict(@RequestParam("file") MultipartFile file) { try { // 保存上传的文件到临时路径 String tempDir = System.getProperty("java.io.tmpdir"); File tempFile = new File(tempDir, file.getOriginalFilename()); file.transferTo(tempFile); // 调用 Python 脚本执行预测 ProcessBuilder pb = new ProcessBuilder( "D:\\Python\\python.exe", "D:/DevCode/商城/Shop-master/shop-springboot/src/main/resources/scripts/image_classifier.py", "predict", tempFile.getAbsolutePath() ); pb.redirectErrorStream(true); // 合并 stdout 和 stderr Process process = pb.start(); BufferedReader reader = new BufferedReader(new InputStreamReader(process.getInputStream())); StringBuilder output = new StringBuilder(); String line; while ((line = reader.readLine()) != null) { output.append(line); } int exitCode = process.waitFor(); if (exitCode == 0) { return ResponseEntity.ok(output.toString().trim()); } else { return ResponseEntity.status(500).body("{\"error\": \"Prediction failed\"}"); } } catch (Exception e) { return ResponseEntity.status(500).body("{\"error\": \"" + e.getMessage() + "\"}"); } } private static final String PYTHON_EXECUTABLE = "D:\\Python\\python.exe"; // 或 "python3" private static final String INFER_SCRIPT_PATH = "D:/DevCode/商城/Shop-master/shop-springboot/src/main/resources/scripts/python-model/infer.py"; @CrossOrigin(origins = "*") @PostMapping("/uploadPython") public Result<?> classifyImage(@RequestParam("file") MultipartFile file) { ObjectMapper mapper = new ObjectMapper(); // 临时保存上传的图片 File tempImage; try { tempImage = File.createTempFile("img_", "_" + file.getOriginalFilename()); file.transferTo(tempImage); } catch (IOException e) { return Result.error("500", "文件保存失败" + e.getMessage()); } try { // 构建命令 ProcessBuilder pb = new ProcessBuilder( PYTHON_EXECUTABLE, INFER_SCRIPT_PATH, tempImage.getAbsolutePath() ); pb.redirectErrorStream(true); // 执行 Process process = pb.start(); // 读取输出 BufferedReader reader = new BufferedReader(new InputStreamReader(process.getInputStream())); StringBuilder output = new StringBuilder(); String line; while ((line = reader.readLine()) != null) { output.append(line); } // 等待完成 boolean finished = process.waitFor(30, TimeUnit.SECONDS); // 超时保护 if (!finished) { process.destroyForcibly(); return Result.error("500", "推理超时"); } reader.close(); // 清理临时文件 tempImage.delete(); // 解析 JSON String jsonOutput = output.toString().trim(); if (jsonOutput.isEmpty()) { return Result.error("500", "Python 无返回结果"); } JsonNode result = mapper.readTree(jsonOutput); System.out.println( result); return Result.success(result); } catch (InterruptedException | JsonProcessingException e) { return Result.error("500", "解析错误"); } catch (Exception e) { return Result.error("500", "系统错误"+e.getMessage()); } } }# python-model/infer.py import tensorflow as tf import numpy as np import sys import os import json from PIL import Image MODEL_PATH = "animal_classifier.h5" IMG_SIZE = 224 CLASS_NAMES = ['cat', 'dog', 'elephant', 'fish'] # 必须与训练时一致! def predict(image_path): if not os.path.exists(image_path): print(json.dumps({"error": f"图片不存在: {image_path}"})) return try: # 加载模型 if not os.path.exists(MODEL_PATH): print(json.dumps({"error": f"模型未找到: {MODEL_PATH}"})) return model = tf.keras.models.load_model(MODEL_PATH) # 加载并预处理图像 img = Image.open(image_path).convert("RGB") img = img.resize((IMG_SIZE, IMG_SIZE)) img_array = np.array(img) / 255.0 img_array = np.expand_dims(img_array, axis=0) # 添加 batch 维度 # 预测 preds = model.predict(img_array, verbose=0) confidence = float(np.max(preds)) label_idx = np.argmax(preds) label = CLASS_NAMES[label_idx] result = { "success": True, "predicted_class": label, "confidence": round(confidence, 4), "all_probabilities": { CLASS_NAMES[i]: round(float(preds[0][i]), 4) for i in range(len(CLASS_NAMES)) } } print(json.dumps(result)) except Exception as e: print(json.dumps({"error": f"预测失败: {str(e)}"})) if __name__ == "__main__": if len(sys.argv) != 2: print(json.dumps({"error": "用法: python infer.py <image_path>"})) else: predict(sys.argv[1])
最新发布
10-22
好,下面是跟在后面的调试信息,帮我描述启动过程等等你能看出的信息。[2025-07-25 15:02:33] [ERROR] ds_handle_method_do():1608 - [DS]not support action get_rtsp_info [2025-07-25 15:02:33] [ERROR] ds_handle():2504 - [DS]Signal handle failed. plug -->DWC_ETH_QOS_setsettings <--DWC_ETH_QOS_setsettings [2025-07-25 15:02:38] [ERROR] system_if_del_gateway_changing MTU from 1500 to 1480 route():267 - [NIFC]Ioctl SIOCDELRT error, ret code is -1, dst =DWC_ETH_QOS_yinit: apb_clk 120000000 Hz 00000000, gateway = 0100a8c0, error info:No such process. Queue0 Tx fifo size 2048, Rx fifo size 2048 Disabled JUMBO pkt Enabled Rx watchdog timer Enabled TSO Disabled Rx Split header mode [2025-07-25 15:02:38] link_status_update():410 - [Information][Network][NIFC]Link status: LINK_DOWN -> LINK_DOWN [2025-07-25 15:02:38] link_status_update():412 - [Information][Network][NIFC]IP: 192.168.0.60, mask: 255.255.255.0, gateway: 192.168.0.1, DNS: 8.8.8.8, 8.8.4.4 [2025-07-25 15:02:38] [ERROR] eap_start():889 - [EAP]g_ieee8021x_enable = 0 timer handle excute times 5, interval 1000 timer handle excute times 5, interval 1000 timer handle excute times 5, interval 1000 timer handle excute times 5, interval 1000 timer handle excute times 5117, interval 1 [2025-07-25 15:02:38] link_status_update():410 - [Information][Network][NIFC]Link status: LINK_DOWN -> LINK_UP [2025-07-25 15:02:38] link_status_update():412 - [Information][Network][NIFC]IP: 192.168.0.60, mask: 255.255.255.0, gateway: 192.168.0.1, DNS: 8.8.8.8, 8.8.4.4 timer handle excute times 3, interval 1 [2025-07-25 15:02:38] [ERROR] eap_reload():912 - [EAP]reload ... g_ether_status:0 phy_status.ether:1 [2025-07-25 15:02:38] [ERROR] ds_handle_method_do():1608 - [DS]not support action get_rtsp_info [2025-07-25 15:02:38] [ERROR] ds_handle():2504 - [DS]Signal handle failed. [2025-07-25 15:02:38] ip_conflict_detect_start():1000 - [Information][Network]Start IP conflict detecting [2025-07-25 15:02:38] [ERROR] ds_handle_method_do():1608 - [DS]not support action get_rtsp_info [2025-07-25 15:02:38] [ERROR] ds_handle():2504 - [DS]Signal handle failed. timer handle excute times 6, interval 1 [2025-07-25 15:02:38] [ERROR] load_user_defined_address():118 - [DIAGNOSE][LTE] add default dns address:8.8.8.8 [2025-07-25 15:02:38] [ERROR] load_user_defined_address():121 - [DIAGNOSE][LTE] add default dns address:www.google.com [2025-07-25 15:02:38] [ERROR] load_user_defined_address():143 - [DIAGNOSE]user defined addrees1:, user define address2:. [2025-07-25 15:02:38] [ERROR] telemetry_post_start():1844 - [telemetry]telemetry post start timer handle excute times 20, interval 1 timer handle excute times 3, interval 1 timer handle excute times 3, interval 1 timer handle excute times 296, interval 1 [2025-07-25 15:02:38] [ERROR] video_stream_load_config():3026 - [AVTS]============stream(0): [2025-07-25 15:02:38] [ERROR] video_stream_load_config():3027 - [AVTS]encode type: H265. [2025-07-25 15:02:38] [ERROR] video_stream_load_config():3028 - [AVTS]resolution: 2560*1440. [2025-07-25 15:02:38] [ERROR] video_stream_load_config():3029 - [AVTS]max resolution: 2560*1440. [2025-07-25 15:02:38] [ERROR] video_stream_load_config():3030 - [AVTS]bitrate type: 2. [2025-07-25 15:02:38] [ERROR] video_stream_load_config():3031 - [AVTS]bitrate: 2560. [2025-07-25 15:02:38] [ERROR] video_stream_load_config():3032 - [AVTS]quality: medium. [2025-07-25 15:02:38] [ERROR] video_stream_load_config():3033 - [AVTS]framerate: 1/25. [2025-07-25 15:02:38] [ERROR] video_stream_load_config():303vprc: enable direct func at in[0] 5 - [AVTS]smart_codec: 1. [2025-07-25 15:02:38] [ERROR]vprc: enable one-buf func at path 0, (3ndr_ref=1) (max_strp=1) video_stream_load_config():3036 - [AVTS]smart_codec_type: 1vcap: enable direct func at out[0] . [2025-07-25 15:02:38] [ERROR] video_stream_load_config():3053 - [AVTS]============================== [2025-07-25 15:02:38] [ERROR] video_stream_load_config():3026 - [AVTS]============stream(1): [2025-07-25 15:02:38] [ERWRN:_isf_vdoprc_oport_do_new() vdoprc0.out[0]! Start single blk mode => blk_id=86001fc0 addr=86002000 ROR] video_stream_load_config():3027 - [AVTS]encode type: H264. [2025-07-25 15:02:38] [ERROR] video_stream_loadERR:nvtanr_setconfig() dB value=0 < 3. Set to 3. _config():3028 - [AVTS]resolution: 848*480. [2025-07-25 15:02:3ERR:nvtanr_setconfig() Bias sensitive value=0 < 1. Set to 1. 8] [ERROR] video_stream_load_config():3029 - [AVTS]m[acap] update cmd 0x2++ ax resolution: 848*576. [2025-07-25 15:02:38] [ERROR]acap] update cmd 0x2-- 0m video_stream_load_config():3030 - [AVTS]bitrate type: 0. [20[acap] update cmd 0x400++ 25-07-25 15:02:38] [ERROR] video_stream_load_config([acap] update cmd 0x400-- ):3031 - [AVTS]bitrate: 512. [2025-07-25 15:02:38] [ERR[acap] update cmd 0x4++ OR] video_stream_load_config():3032 - [AVTS]quality: medium.[acap] update start 1 [2025-07-25 15:02:38] [ERROR] video_stream_load_co[acap] update start 2 nfig():3033 - [AVTS]framerate: 1/25. [2025-07-25 15:02:38] acap] start 1 ;37m[ERROR] video_stream_load_config():3035 - [AVTS]smart_co[acap] start 2 dec: 0. [2025-07-25 15:02:38] [ERROR] video_stream_[acap] start 3 load_config():3036 - [AVTS]smart_codec_type: 0. [2025-07-25 15:[acap] start 4 02:38] [ERROR] video_stream_load_config():3053 - [AV[acap] start 5 TS]============================== [2025-07-25 15:02:38] E_CBMSG_PREVIEWSTABLE_20240320_tp ========== m[ERROR] __audio_stream_init():569 - [MPP]nvt_attr.sample_ra[acap] start 6 te_max:8000 [acap] start 7 [acap] start 15 [acap] start 16 [acap] start 18 [acap] start 20 [acap] start 22 [acap] lb 1 [acap] start 23 [acap] start 24 [acap] start 25 [acap] start 26 [acap] start 27 [acap] start 28 [acap] start 29 [acap] start 30 [acap] update start 3 [acap] update cmd 0x4-- ERR:nvtanr_setconfig() dB value=0 < 3. Set to 3. ERR:nvtanr_setconfig() Bias sensitive value=0 < 1. Set to 1. yanyu uiSetRCMode 2 timer handle excute times 795, interval 1 [2025-07-25 15:02:39] [ERROR] ds_register_op_log_info():720 - [DS]Section harddisk sub_type and func_name is set already. [2025-07-25 15:02:39] [ERROR] processing_linkage_capability():147 - [STM]linkage_capability capability length 30 [2025-07-25 15:02:39] [ERROR] processing_linkage_capability():147 - [STM]linkage_capability capability length 30 [2025-07-25 15:02:39] [ERROR] create_and_init_sem():140 - [STM]creating initial sem... [2025-07-25 15:02:39] [ERROR] init_stm_files_structure():144 - [STM]sizeof(struct stm_file) = 104 [2025-07-25 15:02:39] [ERROR] reload_record_adv_config():2572 - [STM]Set delay time to 10 seconds [2025-07-25 15:02:39] [ERROR] stm_open():2588 - [STM][0] stm_fd is allocated. channel = 0, type = video, flags = 0x1, mode = 0x1 [2025-07-25 15:02:39] [ERROR] data_sync_handle_thread():200 - [STM]register signal handler on thread:0x732e94c0 [2025-07-25 15:02:39] [ERROR] destroy_playback_buf():905 - [STM]Free data_buf 0, size is 0x0 [202ERR:nvtanr_apply() NvtAnr is not opened 5-07-25 15:02:39] [ERROR] destroy_playback_buf():906 - [STM]Free data_buf 1, size is 0x0 [2025-07-25 15:02:39] [ERROR] destroy_playback_buf():905 - [STM]Free data_buf 0, size is 0x0 [2025-07-25 15:02:39] [ERROR] destroy_playback_buf():906 - [STM]Free data_buf 1, size is 0x0 [2025-07-25 15:02:39] [ERROR] destroy_playback_buf():905 - [STM]Free data_buf 0, size is 0x0 [2025-07-25 15:02:39] [ERROR] destroy_playback_buf():906 - [STM]Free data_buf 1, size is 0x0 [2025-07-25 15:02:39] [ERROR] destroy_playback_buf():905 - [STM]Free data_buf 0, size is 0x0 [2025-07-25 15:02:39] [ERROR] destroy_playback_buf():906 - [STM]Free data_buf 1, size is 0x0 [2025-07-25 15:02:39] [ERROR] playback_init():5201 - [STM]playback init over. [2025-07-25 15:02:39] [ERROR] set_auto_delete_timer():713 - [STM]auto delete switch: off [2025-07-25 15:02:39] [ERROR] set_auto_delete_timer():714 - [STM]auto delete record before 7 days [2025-07-25 15:02:39] [ERROR] storage_start():594 - [STM]storage start over. timer handle excute times 15, interval 1 [2025-07-25 15:02:39] [ERROR] ds_convert_init():84 - [DS]ds convert init successed. ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened [2025-07-25 15:02:40] [ERROR] record_stop():4239 - [STM]record stopped ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened DWC_ETH_QOS_adjust_link: start tx/rx DWC_ETH_QOS f02b0000.eth eth0: Link is Up - 100Mbps/Full - flow control rx/tx ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened ERR:nvtanr_apply() NvtAnr is not opened timer handle excute times 3, interval 1000 timer handle excute times 3, interval 1000 [network_send()]:[248]: sendto error: Connection refused [dms_probe_cb()]:[645]: msg mid(1) mlen(196) send failed [network_send()]:[248]: sendto error: Connection refused [dms_sendto()]:[231]: msg mid(36) mlen(0) send failed [2025-07-25 15:02:42] [ERROR] eap_reload():912 - [EAP]reload ... g_ether_status:1 phy_status.ether:0 timer handle excute times 2, interval 1000 timer handle excute times 2, interval 1000 timer handle excute times 2, interval 1000 timer handle excute times 3, interval 1000 [2025-07-25 15:02:42] [ERROR] ds_handle_method_do():1608 - [DS]not support action get_rtsp_info [2025-07-25 15:02:42] [ERROR] ds_handle():2504 - [DS]Signal handle failed. timer handle excute times 3, interval 1000 timer handle excute times 3, interval 1000 timer handle excute times 2819, interval 1 [2025-07-25 15:02:42] [ERROR] rtsp_server_init():224ERR:nvtanr_apply() NvtAnr is not opened 9 - [RTSP]rtsps_support: 0 [2025-07-25 15:02:42] [ERROR] rtsp_server_init():2253 - [RTSP]RTSP Server Port 554 [2025-07-25 15:02:42] [ERROR] rtsp_server_init():2256 - [RTSP]g_tp_rtsp_digest_auth_type: 0 [2025-07-25 15:02:42] [ERROR] rtsp_server_init():2270 - [RTSP]Start RTSP Server on port 554 [2025-07-25 15:02:42] [ERROR] rtsp_server_init():2349 - [RTSP]RTSP Server Init ended [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:0, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:2, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:3, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:4, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:5, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:6, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:7, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:8, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:11, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:12, attach_ringbuffer_enable:0 [2025-07-25 15:02:ERR:nvtanr_apply() NvtAnr is not opened 42] [ERROR] trans_start():2697 - sd_type:14, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:15, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2697 - sd_type:16, attach_ringbuffer_enable:0 [2025-07-25 15:02:42] [ERROR] trans_start():2708 - Try to attach ringbuffer 21000 [2025-07-25 15:02:42] [ERROR] reload_multicast_config():886 - [RTSP][Media Source]: stream type 3 have no multicast config [2025-07-25 15:02:42] [ERROR] reload_multicast_config():886 - [RTSP][Media Source]: stream type 4 have no multicast config [2025-07-25 15:02:42] [ERROR] reload_multicast_config():886 - [RTSP][Media Source]: stream type 5 have no multicast config [2025-07-25 15:02:42] [ERROR] trans_start_loop():1677 - goto send_empty_smart_data [2025-07-25 15:02:42] [ERROR] inetd_register_rtsp_server():2127 - [RTSP]register rtsp server done. [2025-07-25 15:02:42] [ERROR] rtsp_server_start():3548 - [RTSP]trans_attach_to_stream stream_id 7 timer handle excute times 37, interval 1 [2025-07-25 15:02:42] [ERROR] add_rtp_package_info():637 - [RTSP]add package code_type 5 [2025-07-25 15:02:42] [ERROR] init_rtp_package():140 - [RTSP]av_codec = 5 [2025-07-25 15:02:42] [ERROR] init_rtp_packaERR:nvtanr_apply() NvtAnr is not opened ge():185 - [RTSP]jpeg_header width 80 height 45 [2025-07-25 15:02:42] [ERROR] add_rtp_package_info():637 - [RTSP]add package code_type 1 [2025-07-25 15:02:42] [ERROR] init_rtp_package():140 - [RTSP]av_codec = 1 [2025-07-25 15:02:42] [ERROR] RtpAdjustTime():327 - [RTSP]init rtp_base, pts=21285962, utc_ms=1753426962297 [2025-07-25 15:02:42] [ERROR] add_rtp_package_info():637 - [RTSP]add package code_type 0 [2025-07-25 15:02:42] [ERROR] init_rtp_package():140 - [RTSP]av_codec = 0 [2025-07-25 15:02:42] [ERROR] RtpAdjustTime():327 - [RTSP]init rtp_base, pts=21287618, utc_ms=1753426962298 [2025-07-25 15:02:42] [ERROR] add_rtp_package_info():637 - [RTSP]add package code_type 7 [2025-07-25 15:02:42] [ERROR] init_rtp_package():140 - [RTSP]av_codec = 7 [2025-07-25 15:02:42] [ERROR] RtpAdjustTime():327 - [RTSP]init rtp_base, pts=21172583, utc_ms=1753426962183 [2025-07-25 15:02:42] [ERROR] add_rtp_package_info():637 - [RTSP]add package code_type 7 [2025-07-25 15:02:42] [ERROR] init_rtp_package():140 - [RTSP]av_codec = 7 [2025-07-25 15:02:42] [ERROR] trans_register_stream_vtype():2298 - stream srcid = 7 [2025-07-25 15:02:42] [ERROR] RtpAdjustTime():327 - [RTSP]init rtp_base, pts=21172583, utc_ms=1753426962183 [2025-ERR:nvtanr_apply() NvtAnr is not opened 07-25 15:02:42] [ERROR] ds_register_action():824 - [DS]Action [video.force_iframe] has exist. [2025-07-25 15:02:42] link_status_update():410 - [Information][Network][NIFC]Link status: LINK_UP -> LINK_UP [2025-07-25 15:02:42] link_status_update():412 - [Information][Network][NIFC]IP: 192.168.0.60, mask: 255.255.255.0, gateway: 192.168.0.1, DNS: 8.8.8.8, 8.8.4.4 [2025-07-25 15:02:42] [ERROR] eap_reload():912 - [EAP]reload ... g_ether_status:0 phy_status.ether:1 [2025-07-25 15:02:42] [ERROR] ds_handle_method_do():1608 - [DS]not support action get_rtsp_info [2025-07-25 15:02:42] [ERROR] ds_handle():2504 - [DS]Signal handle failed. [2025-07-25 15:02:42] [ERROR] rtsp_reload():3850 - [RTSP]config changes rtsp_reload timer handle excute times 398, interval 1 timer handle excute times 2, interval 1 ERR:nvtanr_apply() NvtAnr is not opened open adc channel 0 error ! open adc channel 1 error ! open adc channel 2 error ! open adc channel 3 error ! open adc channel 0 error ! open adc channel 1 error ! open adc channel 2 error ! open adc channel 3 error ! [2025-07-25 15:02:42] [ERROR] image_3dnr_lv_reconfig():1237 - [CAMERA]image_3dnr_lv_reconfig called [2025-07-25 15:02:42] [ERROR] image_2dnr_lv_reconfig():1313 - [CAMERA]image_2dnr_lv_reconfig called [**************reload isp.cfg **************] [filepath=/tmp/base-files/cfg/day/day_isp.cfg] ERR:nvtanr_apply() NvtAnr is not opened ERR:vos_file_close() vfs_fsync fail, vos_file 0x8283B600 [2025-07-25 15:02:42] [ERROR] image_start():1454 - [IMAGE]Image start over. [2025-07-25 15:02:43] [ERROR] eap_reload():912 - [EAP]reload ... g_ether_status:1 phy_status.ether:0 timer handle excute times 338, interval 1 [2025-07-25 15:02:43] [ERROR] dn_switch_init():1120 - dn_switch,choose IR+WL algorithERR:vos_file_close() vfs_fsync fail, vos_file 0x8283BB00 m [dms_attach_event()]:[782]: callback for mid(28720) exists. [dms_attach_event()]:[782]: callback for mid(28735) exists. timer handle excute times 30, interval 1 timer handle excute times 2, interval 1 [2025-07-25 15:02:43] [ERROR] get_motion_rec_enhance_capability():97 - [RECORD_PLAN]ds_read /smart_analysis/detection fail. [2025-07-25 15:02:43] [ERROR] record_plan_start():385 - [RECORD_PLAN]get_motion_rec_enhance_capability error [2025-07-25 15:02:43] [ERROR] recERR:nvtanr_apply() NvtAnr is not opened ord_plan_start():388 - [RECORD_PLAN]start over, ret[0] [2025-07-25 15:02:43] [ERROR] record_plan_action_callback():1868 - [STM]record_plan new rec type:(motion) timer handle excute times 7, interval 1 [2025-07-25 15:02:43] [ERROR] msg_alarm_update_arming_status():1243 - [MSG_ALARM]alarm type 1, remain 32237s [2025-07-25 15:02:43] [ERROR] msg_alarm_update_arming_status():1285 - [MSG_ALARM]alarm type 0, remain 32237s [2025-07-25 15:02:43] [ERROR] msg_alarm_start():1514 - [MSG_ALARM]msg alarm start over. timer handle excute times 3, interval 1 [2025-07-25 15:02:43] [ERROR] hsr_alarm_start():2616 - [HSR]hsr_alarm start over. timer handle excute times 11, interval 1 [2025-07-25 15:02:43] [ERROR] md_alarm_start():4153 - [MD_ALARM]md alarm start over. timer handle excute times 6, interval 1 [2025-07-25 15:02:43] [ERROR] od_alarm_start():749 - [OD_ALARM]od alarm start over. timer handle excute times 2, interval 1 [2025-07-25 15:02:43] [ERROR] id_alarm_start():4112 - [ID]id_alarm start over. timer handle excute times 16, interval 1 [2025-07-25 15:02:43] [ERROR] cd_alarm_start():3109 - [CD]cd_alarm start over. timer handle excute times 10, interval 1 [2025-07-25 15:02:43] [ERROR] vd_alarm_start():2448 - [VD_ALARM]vd_alarm start over. timer handle excute times 4, interval 1 [2025-07-25 15:02:43] [ERROR] er_alarm_start():3323 - [ER]er_alarm start over. timer handle excute times 17, interval 1 [2025-07-25 15:02:43] [ERROR] lr_alarm_start():3328 - [LR]lr_alarm start over. timer handle excute times 4, interval 1 [2025-07-25 15:02:43] [ERROR] aod_alarm_start():2918 - [AOD]aod_alarm start over. timer handle excute times 20, interval 1 [2025-07-25 15:02:43] [ERROR] aad_alarm_start():1116 - [AAD]aad_alarm start over.ERR:vos_file_open() open [/mnt/app/isp/day_isp_dpc.bin] failed, ret -2 timer handle excute times 3, interval 1 [2025-07-25 15:02:43] [ERROR] hd_alarm_start():3029 - [HD]hd_alarm start over. timer handle excute times 8, interval 1 [2025-07-25 15:02:43] [ERROR] sc_alarm_start():975 - [SC]sc_alarm start over. timer handle excute times 4, interval 1 [2025-07-25 15:02:43] [ERROR] ams_start():557 - [AMS] ams version: 2.0.0.4348-b9f4532-dirty timer handle excute times 5, interval 1 [dms_detach_event()]:[850]: detach mERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs.bin] failed, ret -2 sg[20486] handler[0x3b0fdc] [dms_detach_event()]:[850]: detach msg[20487] handler[0x3b0e90] timer handle excute times 10, interval 1 [2025-07-25 15:02:43] [ERROR] p2p_start():34ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_0.bin] failed, ret -2 - [P2P]p2p module started. version: 2.0 p2p do punch start. timer handle excute times 2, interval 1 timer handle excute times 3, interval 1 error was 0 ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_1.bin] failed, ret -2 over. [2025-07-25 15:02:43] aecns_init():924 - [Alarm][AMS] [TPAECNS] init ok! [2025-07-25 15:02:43] link_status_update():410 - [Information][Network][NIFC]Link status: LINK_UP -> LINK_UP [2025-07-25 15:02:43] [ERROR] dla_context_init_dev():1139 -ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_2.bin] failed, ret -2 [AMS] DLA read TP_MODE_SWITCH failed [2025-07-25 15:02:43] [ERROR] dla_load_alg_on_cfg_dev():2154 - [AMS] read PTD_DETECT failed, PET moduERR:vos_file_open() open [/mnt/app/isp/day_isp_lut2d.bin] failed, ret -2 le config file not found. [2025-07-25 15:02:43] link_status_update():412 - [Information][Network][NIFC]IP: 192.168.0.60, mask: 255.255.255.0, gateway: 192.168.0.1, DNS: 8.8.8.8, 8.8.4.4 [2025-07-25 15:02:43] [ERROR] eap_reload():912 - [EAP]reload ... g_ether_status:0 phy_status.ether:1 [2025-07-25 15:02:43] [ERROR] ds_handle_method_do():1608 - [DS]nERR:vos_file_close() vfs_fsync fail, vos_file 0x83939400 ot support action get_rtsp_info [2025-07-25 15:02:43] [ERROR] ds_handle():2504 - [DS]Signal handle failed. [2025-07-25 15:02:43] [ERROR] dla_ctx_init():138 - [2025-07-25 15:02:43] [ERROR] rtsp_reload():3850 - [RTSP]config changes rtsp_reload [2025-07-25 15:02:43] [ERROR] rtmp_reload():1213 - [RTMP]rtmp reload [2025-07-25 15:02:43] [ERROR] rtmp_reconnect():1182 - [RTMP]stop stream flag:0 pClient->lte_stop_flag:0 timer handle excute times 218, interval 1 [2025-07-25 15:02:43] [ERROR] ftp_record_plan_start():767 - [FTP_RECORD_PLAN]start over, ret[0] [2025-07-25 15:02:43] [ERROR] ftp_record_plan_action_callback():849 - [FTP]ftp_record_plan new rec type:(null) [2025-07-25 15:02:43] [ERROR] openapi_load():672 - [OPENAPI]/opeERR:vos_file_close() vfs_fsync fail, vos_file 0x82945C00 napi/server disable [2025-07-25 15:02:43] [ERROR] openapi_start():1067 - [OPENAPI]openapi disable [libdla] libdla version: 2.0.0.4348-b9f4532-dirty [2025-07-25 15:02:43] [ERROR] update_server_info():2281 - [SMTP]no need to send email timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 update_core_dump_config()-57 - [DEBUG_TOOLS]enable coredump [**************reload isp.cfg **************] [filepath=/tmp/base-files/cfg/day/day_isp.cfg] enable_core_dump()-42 - [DEBUG_TOOLS]new core dump file size: cur 4294967295, max 4294967295 Monitor: receive NVMP_NO_RESTART. (uptime:23) debug_tools_coredump_start()-100 - [DEBUG_TOOLS]USAGE:input "killall -10 main" to enable coredump timer handle excute times 172, interval 1 timer handle excute times 3, interval 1 timer handle excute times 2, interval 1 [2025-ERR:vos_file_close() vfs_fsync fail, vos_file 0x82945D00 07-25 15:02:43] [ERROR] mb_start():4958 - [MB]ringbuffer_attach error:type:0, rb_id:10201 [2025-07-25 15:02:43] [ERROR] mb_start():4958 - [MB]ringbuffer_attach error:type:1, rb_id:10202 [2025-07-25 15:02:43] [ERROR] mb_start():4958 - [MB]ringbuffer_attach error:type:2, rb_id:10200 [2025-07-25 15:02:43] [ERROR] mb_start():4958 - [MB]ringbuffer_attach error:type:3, rb_id:10206 [2025-07-25 15:02:43] [ERROR] mb_update_target_valid():1066 - [MB]mb_update_target_valid[0] [2025-07-25 15:02:43] [ERROR] mb_update_target_valid():1066 - [MB]mb_update_target_valid[0] [2025-07-25 15:02:43] [ERROR] mb_update_target_valid():1066 - [MB]mb_update_target_valid[0] ai - proc[0].trace = 00000000 [2025-07-25 15:02:43] [ERROR] mb_update_target_valid():1066 - [MB]mb_update_target_valid[0] [2025-07-25 15:02:43] [ERROR] qd_update_mb_enabled():3243 - [MB]dsusb 1-1: USB disconnect, device number 2 read /queue_detection/detection error [2025-07-25 15:02:43] [ERROR] mb_start():5025 - [MB]mb start over. timer harndis_host 1-1:1.0 usb0: unregister 'rndis_host' usb-ehci_hcd-1, RNDIS device ndle excute times 23, interval 1 [2025-07-25 15:02:43] [ERROR] dla_get_models():94 - [AMS] rotate type: 3, model list :0, 1, 2, 3 timer handle excute times 15, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 [2025-07-25 15:02:43] [ERROR] update_server_info():322 - [REPORT_EVENT]read report_server_list failed [2025-07-25 15:02:43] [ERROR] update_server_info():322 - [REPORT_EVENT]read report_server_list failed [2025-07-25 15:02:43] [ERROR] update_server_info():322 - [REPORT_EVoption1 ttyUSB0: GSM modem (1-port) converter now disconnected from ttyUSB0 ENT]read report_server_list failed timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, inteoption 1-1:1.2: device disconnected rval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute option1 ttyUSB1: GSM modem (1-port) converter now disconnected from ttyUSB1 times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 3, interval 1 [2025-07-25 15:02:43] [ERROR] _getsize_model():141 - timer handle excute times 2, interval 1 [libdla] get bin(/etc/plugins/obj_detection_fd.bin) size fail [2025-07-25 15:02:43] [ERROR] _getsize_model():141 - [libdla] ption 1-1:1.3: device disconnected 1mget bin(/etc/plugins/obj_detection_fd_90.bin) size fail [2025-07-25 15:02:43] [ERROR] _getsize_model():141 - [libdla] get bin(/etc/plugins/obj_detection_fdERR:vos_file_open() open [/mnt/app/isp/day_isp_dpc.bin] failed, ret -2 _90.bin) size fail [2025-07-25 15:02:43] [ERROR] hw_network_load():304 - [libdla] ***********option1 ttyUSB2: GSM modem (1-port) converter now disconnected from ttyUSB2 ***** max_blk_size:2948576 ***************** timer handle eERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs.bin] failed, ret -2 xcute times 3, interval 1 option 1-1:1.4: device disconnected ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_0.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_1.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_2.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_lut2d.bin] failed, ret -2 ERR:vos_file_close() vfs_fsync fail, vos_file 0x83939400 nvtim_patch_usbhc enter error was 0 [2025-07-25 15:02:44]lte_start():12016 - [LTE]g_delta_total_flow:0 [2025-07-25 15:02:44] [ERROR] image_3dnr_lv_reconfig():1237 - [CAMERA]image_3dnr_lv_reconfig called [2025-07-25 15:02:44] [ERROR] image_2dnr_lv_reconfig():1313 - [CAMERA]image_2dnr_lv_reconfig called timer handle excute times 1009, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 5, interval 1 timer handle excute times 2, interval 1 timer handle excute times 3, interval 1 timer handle excute times 2, interval 1 timer handle excute times 43, interval 1 timer handle excute times 2, interval 1 timer handle excute times 8, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 [2025-07-25 15:02:44] delay_setting_light_timer_handl():3185 - [Information][Camera]Switch to DAY, WTL OFF [2025-07-25 15:02:44] delay_setting_light_timer_handl():3190 - [Information][Camera]Switch to DAY, IR OFF [**************reload isp.cfg **************] [filepath=/tmp/base-files/cfg/day/day_isp.cfg] timer handle excute times 7, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 timer handle excute times 2, interval 1 [2025-07-25 15:02:44] [ERROR] set_time_date_format():4448 - [CAMERA]change time format:0, change date format:0 [2025-07-25 15:02:44] [ERROR] osd_cal_max_font_size():1824 - [CAMERA]get max main resolusion: 2560, calc main font size:80 , get max minor resolusion: 848, calc minor max font size: 28 [2025-07-25 15:02:44] [ERROR] osd_init_bitmap():2034 - [CAMERA]init bitmap main size:72, char_size:72, res:2560 [2025-07-25 15:02:44] [ERROR] regenerate_mem_font():4029 - [CAMERA]regenerate_mem_font called [2025-07-25 15:02:44] [ERROR] fill_mem_font_head():4101 - [CAMERA]after default write offset:132, data size:100 [2025-07-25 15:02:44] [ERROR] fill_mem_font_head():4177 - [CERR:vos_file_close() vfs_fsync fail, vos_file 0x84241900 AMERA]g write offset:7596 [2025-07-25 15:02:44] [ERROR] camera_osd_start():3932 - [CAMERA]cated string:APSatunMoTehWdFri1234567890 .-:ZVIGI C540-4G 1.20_D936, g_is_font_changed:1 [2025-07-25 15:02:44] [ERROR] osd_cal_max_font_size():1824 - [CAMERA]get max main resolusion: 2560, calc main font size:80 , get max minor resolusion: 848, calc minor max font size: 28 [2025-07-25 15:02:44] [ERROR] get_stampsize():2656 - [CAMERA]chn:0, osd mem size:0x32100 [2025-07-25 15:02:44] [ERROR] set_glyph_info():2674 - [CAMERA]unicode:32, num of vertex:0 [2025-07-25 15:02:44] [ERROR] set_glyph_info():2681 - [CAMERA]vertex pts is NULL [2025-07-25ERR:vos_file_close() vfs_fsync fail, vos_file 0x82945E00 15:02:44] [ERROR] set_glyph_info():2674 - [CAMERA]unicode:32, num of vertex:0 [2025-07-25 15:02:44] [ERROR] set_glyph_info():2681 - [CAMERA]vertex pts is NULL [2025-07-25 15:02:45] [ERROR] osd_cal_max_font_size():1824 - [CAMERA]get max main resolusion: 2560, calc main font size:80 , get max minor resolusion: 848, calc minor max font size: 28 [2025-07-25 15:02:45] [ERROR] get_stampsize():2656 - [CAMERA]chn:0, osd mem size:0x32100 [2025-07-25 15:02:45] [ERROR] set_time_date_format():4448 - [CAMERA]change time format:0, change date format:0 [2025-07-25 15:02:45] [ERROR] osd_cal_max_font_size():1824 - [CAMERA]get max main resolusion: 2560, calc main font size:80 , get max minor resolusion: 848, calc minor max font size: 28 [2025-07-25 15:02:45] [ERROR] osd_init_bitmap():1997 - [CAMERA]minor res:848, main size:72, height:24, width:24 [2025-07-25 15:02:45] [ERROR] camera_osd_start():3932 - [CAMERA]cated string:APSatunMoTehWdFri1234567890 .-:ZVIGI C540-4G 1.20_D936, g_is_font_changed:0 [2025-07-25 15:02:45] [ERROR] osd_cal_max_font_size():1824 - [CAMERA]get max main resolusion: 2560, calc main font size:80 , get max minor resolusion: 848, calc minor max font size: 28 [2025-07-25 15:02:45] [ERROR] get_stampsize():2670 - [CAMERA]chn:1, osd mem size:0x6300 [2025-07-25 15:02:45] [ERROR] osd_cal_max_font_size():1824 - [CAMERA]get max main resolusion: 2560, calc main font size:80 , get max minor resolusion: 848, calc minor max font size: 28 [2025-07-25 15:02:45] [ERROR] get_stampsize():2670 - [CAMERA]chn:1, osd mem size:0x6300 [2025-07-25 15:02:45] [ERROR] osd_start():2285 - [OSD]Osd start over. ERR:vos_file_open() open [/mnt/app/isp/day_isp_dpc.bin] failed, ret -2 error was 0 ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_0.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_1.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_2.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_lut2d.bin] failed, ret -2 ERR:vos_file_close() vfs_fsync fail, vos_file 0x83939700 [**************reload isp.cfg **************] [filepath=/tmp/base-files/cfg/day/day_isp.cfg] ERR:vos_file_close() vfs_fsync fail, vos_file 0x83939A00 [2025-07-25 15:02:45] [ERROR] udp_send_dev_advertisement():1794 - [IPCD]udp data send error ret = -1 ERR:vos_file_close() vfs_fsync fail, vos_file 0x83909500 usb 1-1: new high-speed USB device number 3 using ehci-nvtivot [2025-07-25 15:02:45] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idVendor. [2025-07-25 15:02:45] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idProduct. [2025-07-25 15:02:45] [ERROR] usbdevice_detect():663 - [LTE]find USB device fail ERR:vos_file_open() open [/mnt/app/isp/day_isp_dpc.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_0.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_1.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_ecs_auto_2.bin] failed, ret -2 ERR:vos_file_open() open [/mnt/app/isp/day_isp_lut2d.bin] failed, ret -2 usb 1-1: New USB device found, idVendor=2ecc, idProduct=3001, bcdDevice= 0.00 ERR:vos_file_close() vfs_fsync fail, vos_file 0x82945B00 usb 1-1: New USB device strings: Mfr=3, Product=2, SerialNumber=0 usb 1-1: Product: WUKONG usb 1-1: Manufacturer: MARVELL usb 1-1: USB disconnect, device number 3 error was 0 [2025-07-25 15:02:46] [ERROR] image_3dnr_lv_reconfignvtim_patch_usbhc enter ():1237 - [CAMERA]image_3dnr_lv_reconfig called [2025-07-25 15:02:46] [ERROR] image_2dnr_lv_reconfig():1313 - [CAMERA]image_2dnr_lv_reconfig called [2025-07-25 15:02:46] delay_setting_light_timer_handl():3185 - [Information][Camera]Switch to DAY, WTL OFF [2025-07-25 15:02:46] delay_setting_light_timer_handl():3190 - [Information][Camera]Switch to DAY, IR OFF [2025-07-25 15:02:46] [ERROR] hw_network_load():462 - [libdla] ************* load network 7 success [2025-07-25 15:02:46] [ERROR] smart_box_reload_det():309 - [AMS] read PTD_DETECT failed, PTD module config file not found. [2025-07-25 15:02:46] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idVendor. [2025-07-25 15:02:46] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idProduct. [2025-07-25 15:02:46] [ERROR] usbdevice_detect():663 - [LTE]find USB device fail [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. error was 0 [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idVendor. [2025-07-25 15:02:47] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idProduct. [2025-07-25 15:02:47] [ERROR] usbdevice_detect():663 - [LTE]find USB device fail [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_vd_result():506 - [AMS] mark_motion_vd_result: empty fg! [2025-07-25 15:02:47] [ERROR] vd_dla_process():931 - [AMS] mark_motion_vd_result error. [2025-07-25 15:02:47] [ERROR] vd_dla_trigger():1079 - [AMS] vd_dla_process failed. [2025-07-25 15:02:47] [ERROR] mark_motion_pd_result():495 - [AMS] mark_motion_pd_result: empty fg! [2025-07-25 15:02:47] [ERROR] pd_dla_process():751 - [AMS] mark_motion_pd_result error. [2025-07-25 15:02:47] [ERROR] pd_dla_trigger():1064 - [AMS] pd_dla_process failed. [2025-07-25 15:02:48] [ERROR] ftp_record_start_delay():29 - [FTP]ftp_record_start_delay error was 0 [2025-07-25 15:02:48] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idVendor. [2025-07-25 15:02:48] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idProduct. [2025-07-25 15:02:48] [ERROR] usbdevice_detect():663 - [LTE]find USB device fail error was 0 [2025-07-25 15:02:49] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idVendor. [2025-07-25 15:02:49] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idProduct. [2025-07-25 15:02:49] [ERROR] usbdevice_detect():663 - [LTE]find USB device fail error was 0 [2025-07-25 15:02:50] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idVendor. [2025-07-25 15:02:50] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idProduct. [2025-07-25 15:02:50] [ERROR] usbdevice_detect():663 - [LTE]find USB device fail error was 0 [2025-07-25 15:02:51] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idVendor. [2025-07-25 15:02:51] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idProduct. [2025-07-25 15:02:51] [ERROR] usbdevice_detect():663 - [LTE]find USB device fail [2025-07-25 15:02:52] md_alarm_start_effect():2769 - [Alarm][Motion Detection]Motion detection start [2025-07-25 15:02:52] [ERROR] change_record_status():1238 - [STM]rec status change: pre => recording error was 0 [2025-07-25 15:02:52] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idVendor. [2025-07-25 15:02:52] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idProduct. [2025-07-25 15:02:52] [ERROR] usbdevice_detect():663 - [LTE]find USB device fail error was 0 [2025-07-25 15:02:53] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idVendor. [2025-07-25 15:02:53] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idProduct. [2025-07-25 15:02:53] [ERROR] usbdevice_detect():663 - [LTE]find USB device fail error was 0 usb 1-1: new high-speed USB device number 4 using ehci-nvtivot [2025-07-25 15:02:54] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idVendor. [2025-07-25 15:02:54] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-0:1.0/idProduct. [2025-07-25 15:02:54] [ERROR] usbdevice_detect():663 - [LTE]find USB device fail usb 1-1: New USB device found, idVendor=2c7c, idProduct=6005, bcdDevice= 3.18 usb 1-1: New USB device strings: Mfr=1, Product=2, SerialNumber=3 usb 1-1: Product: Android usb 1-1: Manufacturer: Android usb 1-1: SerialNumber: 0000 rndis_host 1-1:1.0 usb0: register 'rndis_host' at usb-ehci_hcd-1, RNDIS device, 5a:e6:53:d5:1a:0c option 1-1:1.2: GSM modem (1-port) converter detected usb 1-1: GSM modem (1-port) converter now attached to ttyUSB0 option 1-1:1.3: GSM modem (1-port) converter detected usb 1-1: GSM modem (1-port) converter now attached to ttyUSB1 option 1-1:1.4: GSM modem (1-port) converter detected usb 1-1: GSM modem (1-port) converter now attached to ttyUSB2 [2025-07-25 15:02:55] md_alarm_end_effect():2837 - [Alarm][Motion Detection]Motion detection end error was 0 [2025-07-25 15:02:55] [ERROR] __enlarge_udesc_list():703 - [AVDM]avdm0 udesc list enlarge 200 to 600 [2025-07-25 15:02:55] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-1:1.3/idVendor. [2025-07-25 15:02:55] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-1:1.3/idProduct. [2025-07-25 15:02:55] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-1:1.1/idVendor. [2025-07-25 15:02:55] [ERROR] get_value_from_file():600 - [LTE]Fail to open /sys/bus/usb/devices/1-1:1.1/idProduct. [2025-07-25 15:02:55] [ERROR] usbdevice_detect():651 - [LTE]Find /sys/bus/usb/devices/1-1 idVendor=0x2c7c idProduct=0x6005 [2025-07-25 15:02:55]lte_timer_handle():10906 - [LTE]Now create lte_main_loop thread. [2025-07-25 15:02:55]openport():58 - [LTE]Starting serial communication process. [2025-07-25 15:02:55] [ERROR] openport():68 - [LTE]Open /dev/ttyUSB1 successfully. [2025-07-25 15:02:55]config_AT_port():3375 - [LTE]opened the device /dev/ttyUSB1 [2025-07-25 15:02:55]setport():207 - [LTE]Comport Init Successfully ...... [2025-07-25 15:02:55] [ERROR] readport():451 - [LTE]urc not found, recv [RDY ] [2025-07-25 15:02:55] [ERROR] readport():451 - [LTE]urc not found, recv [+CFUN: 1 ] [2025-07-25 15:02:55] [ERROR] readport():451 - [LTE]urc not found, recv [+QSIMSTAT: 1,1 ] error was 0 [2025-07-25 15:02:57] [ERROR] config_AT_port():3396 - [LTE]Serial Port:/dev/ttyUSB1 does not repsonse! wait and try later error was 0 [2025-07-25 15:02:58]openport():58 - [LTE]Starting serial communication process. [2025-07-25 15:02:58] [ERROR] openport():68 - [LTE]Open /dev/ttyUSB1 successfully. [2025-07-25 15:02:58]config_AT_port():3375 - [LTE]opened the device /dev/ttyUSB1 [2025-07-25 15:02:58]setport():207 - [LTE]Comport Init Successfully ...... [2025-07-25 15:02:58] [ERROR] readport():451 - [LTE]urc not found, recv [AT ] [2025-07-25 15:02:58]config_AT_port():3404 - [LTE]Find AT_PROT:/dev/ttyUSB1 success [2025-07-25 15:02:58]openport():58 - [LTE]Starting serial communication process. [2025-07-25 15:02:58] [ERROR] openport():68 - [LTE]Open /dev/ttyUSB1 successfully. [2025-07-25 15:02:58]setport():207 - [LTE]Comport Init Successfully ...... [2025-07-25 15:02:58]lte_start_main_loop():6111 - [LTE]lte_main_loop thread start running. [2025-07-25 15:02:58]update_led_state():11558 - [LTE]LED send led event 20 [2025-07-25 15:02:58] [ERROR] lte_start_main_loop():6136 - [LTE]disable echo characters error was 0 [2025-07-25 15:02:58] [ERROR] readport():451 - [LTE]urc not found, recv [ATE0 ] [2025-07-25 15:02:58] [ERROR] readport():451 - [LTE]urc not found, recv [+QIND: SMS DONE ] [2025-07-25 15:02:58] [ERROR] lte_start_main_loop():6146 - [LTE]set ATE0 success. [2025-07-25 15:02:58]lte_start_main_loop():6155 - [LTE]Check if the 4G module is in network interface mode. [2025-07-25 15:02:58]lte_start_main_loop():6186 - [LTE]AT+QCFG="nat" return +QCFG: "nat",1 [2025-07-25 15:02:58]lte_start_main_loop():6218 - [LTE]Check if the 4G module is in rndis mode. [2025-07-25 15:02:58]lte_start_main_loop():6249 - [LTE]AT+QCFG="usbnet" return +QCFG: "usbnet",3 error was 0 error was 0 [2025-07-25 15:03:00]lte_start_main_loop():6643 - [LTE]Main loop init 4G ... [2025-07-25 15:03:00] [ERROR] ping_start():508 - [DIAGNOSE]start ping:8.8.8.8 [2025-07-25 15:03:00] [ERROR] ping_start():508 - [DIAGNOSE]start ping:www.google.com [2025-07-25 15:03:00]clearProfileForESM():3557 - [LTE]clear Profile For default success! [2025-07-25 15:03:00] [ERROR] readport():451 - [LTE]urc not found, recv [+QUSIM: 1 ] [2025-07-25 15:03:00] [ERROR] set_module_properties():3881 - [LTE]SET DDR CLK to 1066 [2025-07-25 15:03:01] [ERROR] set_module_properties():3902 - [LTE]Close SIM write [2025-07-25 15:03:01]waiting_for_sim_det_ctrl():3683 - [LTE]ignored SimDetection set! error was 0 [2025-07-25 15:03:01] md_alarm_start_effect():2769 - [Alarm][Motion Detection]Motion detection start [2025-07-25 15:03:01]update_module_info_param():2931 - [LTE]update_module_info_param now! [2025-07-25 15:03:01]update_module_info_param():2934 - [LTE][vender]Quectel [2025-07-25 15:03:01]update_module_info_param():2935 - [LTE][product]EC200A [2025-07-25 15:03:01]update_module_info_param():2936 - [LTE][revision]EC200AELLAR01A05M16_TP [2025-07-25 15:03:01]update_module_info_param():2937 - [LTE][version]05 [2025-07-25 15:03:01]update_module_info_param():2938 - [LTE][imei]864042070769390 [2025-07-25 15:03:01]update_module_info_param():2967 - [LTE]update_module_info_param now! [2025-07-25 15:03:01] [ERROR] handle_4g_internet_status():9733 - [LTE]4G internet not connected [2025-07-25 15:03:01]lte_start_main_loop():6788 - [LTE]Main loop sim check... [2025-07-25 15:03:01]lte_start_main_loop():6806 - [LTE]start do sim check [2025-07-25 15:03:01] [ERROR] get_sim_card_status():11292 - [LTE]get_sim_card_status cpinLine:+CPIN: READY [2025-07-25 15:03:01]get_cme_err_code():3491 - [LTE]get_cme_err_code cme_err_reply:+CPIN: READY [2025-07-25 15:03:01]get_sim_card_status():11391 - [LTE]SIM card status changed(1 -> 2) [2025-07-25 15:03:01]iccid_manager_handle():11209 - [LTE]4G not connected. [2025-07-25 15:03:01]iccid_manager_handle():11232 - [LTE]lte iccid updated ... [2025-07-25 15:03:01]handle_iccid():11158 - [LTE]SIM_QCCID:89861124640282341661 [2025-07-25 15:03:01] [ERROR] iccid_manager_handle():11251 - [LTE]sim card iccid change. [2025-07-25 15:03:01] [ERROR] iccid_manager_handle():11252 - [LTE]update data_traffic_count and data_traffic_count_start_time. [2025-07-25 15:03:01]iccid_manager_handle():11267 - [LTE]update_iccid now! [2025-07-25 15:03:01] [ERROR] ds_pack_config():978 - [DS]encrypt_key_type:0 [2025-07-25 15:03:02]update_module_info_param():2909 - [LTE]lte module_info is already updated. [2025-07-25 15:03:02] [ERROR] readport():451 - [LTE]urc not found, recv [+QIND: SMS DONE ] [2025-07-25 15:03:02] md_alarm_start_effect():2769 - [Alarm][Motion Detection]Motion detection start [2025-07-25 15:03:02] [ERROR] urc_CGREG():214 - [LTE]ps_net_info.pcid:9B21, ps_net_info.cellID:858BD03 [2025-07-25 15:03:02] [ERROR] readport():451 - [LTE]urc not found, recv [+CGEV: EPS PDN ACT 1 ] error was 0 [2025-07-25 15:03:02] md_alarm_end_effect():2837 - [Alarm][Motion Detection]Motion detection end [2025-07-25 15:03:02] [ERROR] get_cimi_info_param():2894 - [LTE]imsi:460115279906397 [2025-07-25 15:03:02]update_module_info_param():2957 - [LTE][imsi]460115279906397 [2025-07-25 15:03:02]update_module_info_param():2967 - [LTE]update_module_info_param now! [2025-07-25 15:03:02] [ERROR] handle_4g_internet_status():9733 - [LTE]4G internet not connected [2025-07-25 15:03:02]lte_start_main_loop():7293 - [LTE]Main loop search for internet. [2025-07-25 15:03:02]update_led_state():11558 - [LTE]LED send led event 20 [2025-07-25 15:03:02] [ERROR] readport():446 - [LTE]urc 0x10000019 have no proc fun, [+CTZV: "+32" ] [2025-07-25 15:03:02] [ERROR] lte_start_main_loop():7423 - [LTE]err = 5, cgreg_status:1 [2025-07-25 15:03:03] [ERROR] urc_CNEC_ESM_GSM():453 - [LTE]err_code:51,err_cid:8 [2025-07-25 15:03:03] [ERROR] urc_CNEC_ESM_GSM():454 - [LTE]Network side rejection occurred [2025-07-25 15:03:03] [ERROR] urc_CNEC_ESM_GSM():470 - [LTE]no need to Switch profile [2025-07-25 15:03:03] [ERROR] readport():451 - [LTE]urc not found, recv [+CGEV: ME PDN ACT 8,1 ] [2025-07-25 15:03:03] [ERROR] lte_start_main_loop():7470 - [LTE]err = 2, cereg_status:1 [2025-07-25 15:03:03] [ERROR] lte_start_main_loop():7514 - [LTE]ps network register successs, get profile now. [2025-07-25 15:03:03] [ERROR] get_sim_spn_info():524 - [LTE]AT_Response_buf:^SPN: 1,1,▒N-V▒u5O▒ [2025-07-25 15:03:03] [ERROR] get_sim_spn_info():528 - [LTE]short_eons:▒N-V▒u5O▒ [2025-07-25 15:03:03] [ERROR] get_sim_spn_info():531 - [LTE]SIM short_spn is ▒N-V▒u5O▒, coding =1 [2025-07-25 15:03:03] [ERROR] get_sim_gid1_info():474 - [LTE]Sim GID1 Response Error! [2025-07-25 15:03:03] [ERROR] get_sim_gid1_info():478 - [LTE]SIM GID1 is ff len = 0 [2025-07-25 15:03:03] link_up_timer_handle():1993 - [Exception][Network][NIFC]link up timeout! error was 0 [2025-07-25 15:03:03] [ERROR] get_profile():4174 - [LTE]isp:46011. [2025-07-25 15:03:03] [ERROR] get_profile():4177 - [LTE]mcc:460 mnc:11. [2025-07-25 15:03:03] [ERROR] __enlarge_udesc_list():703 - [AVDM]avdm0 udesc list enlarge 200 to 800 [2025-07-25 15:03:03] [ERROR] get_profile():4264 - [LTE]g_apn_conf_index:0 [2025-07-25 15:03:03] [ERROR] get_profile():4266 - [LTE]g_profile_pos:2310 [2025-07-25 15:03:03] [ERROR] get_profile():4281 - [LTE]###########is mvno ERROR######### [2025-07-25 15:03:03] get_profile():4343 - [Information][Network][4G]cur_profile apn:ctnet ip_version:2 username: password: authentication:0 pkg_name:China Telecom (4G) [2025-07-25 15:03:03] [ERROR] get_profile():4354 - [LTE]g_apn_conf_index:1868 [2025-07-25 15:03:03] [ERROR] get_profile():4355 - [LTE]g_profile_pos:1868 [2025-07-25 15:03:03]get_default_bearer_apn():4666 - [LTE]AT*CGDFLT? return:*CGDFLT: "IPV4V6","",0,1,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0. [2025-07-25 15:03:03] [ERROR] get_default_bearer_apn():4668 - [LTE]pdp_type_str:IPV4V6, apn_str: [2025-07-25 15:03:03] [ERROR] get_default_bearer_apn():4672 - [LTE]parse AT*CGDFLT success. [2025-07-25 15:03:03]lte_start_main_loop():7561 - [LTE]isEsmAPN:0, g_prof_config.apn:ctnet,apn_str:,g_prof_config.ip_ver:2,pdp_type:2 [2025-07-25 15:03:03] [ERROR] lte_start_main_loop():7572 - [LTE]ps network register successs, go to APN_SET status. [2025-07-25 15:03:03]iccid_manager_handle():11209 - [LTE]4G not connected. [2025-07-25 15:03:03]iccid_manager_handle():11232 - [LTE]lte iccid updated ... [2025-07-25 15:03:03] [ERROR] iccid_manager_handle():11236 - [LTE]lte module_info is already updated. [2025-07-25 15:03:03] [ERROR] iccid_manager_handle():11237 - [LTE]no need to upload iccid:89861124640282341661. [2025-07-25 15:03:03]lte_start_main_loop():7796 - [LTE]Main loop apn set ...(1) [2025-07-25 15:03:03] [ERROR] handle_4g_internet_status():9733 - [LTE]4G internet not connected [2025-07-25 15:03:03] [ERROR] long_connect_reload():4392 - [LONG_CONNECT]4g internet link_status change to :2 [2025-07-25 15:03:03]lte_start_main_loop():7841 - [LTE]Will not set user apn! [2025-07-25 15:03:03] [ERROR] try_to_set_apn():9463 - [LTE]cmd:AT+CGDCONT=2,"IPV4V6","ctnet" [2025-07-25 15:03:03] [ERROR] remote_log_reload():434 - [REMOTE_LOG]4g internet link_status change to :2 [2025-07-25 15:03:03] [ERROR] ddns_reload():693 - [DDNS]LTE mode: link_status change to :2 error was 0 [2025-07-25 15:03:04] [ERROR] try_to_set_apn():9478 - [LTE]cmd:AT*AUTHREQ=2,0,"","" [2025-07-25 15:03:04] [ERROR] diagnose_msg_send():846 - [DIAGNOSE]send lte model diagnose msg out id:2 [2025-07-25 15:03:04] [ERROR] diagnose_result_save_config():1055 - [DIAGNOSE]internet status has changed: [0] ===> [4] [2025-07-25 15:03:04]lte_net_detection_callback():11759 - [LTE]4G internet not connected [2025-07-25 15:03:04]lte_start_main_loop():7850 - [LTE]set default apn success [2025-07-25 15:03:04] [ERROR] lte_start_main_loop():7888 - [LTE]4G model start attach. [2025-07-25 15:03:04] start_attach_process():795 - [Information][Network][4G]4G model attach successful error was 0 [2025-07-25 15:03:05] [ERROR] lte_status_check_process():10381 - [LTE]sig_lvl:4 [2025-07-25 15:03:05] [ERROR] get_flow_bytes():4076 - [LTE]read_buf:0,data:0 [2025-07-25 15:03:05] [ERROR] get_flow_bytes():4076 - [LTE]read_buf:0,data:0 [2025-07-25 15:03:05]lte_start_main_loop():7925 - [LTE]Main loop start dial... [2025-07-25 15:03:05] [ERROR] udp_send_dev_advertisement():1794 - [IPCD]udp data send error ret = -1 error was 0 [2025-07-25 15:03:06] [ERROR] smart_box_reload_det():309 - [AMS] read PTD_DETECT failed, PTD module config file not found. [2025-07-25 15:03:07] [ERROR] soap_update_global():126 - [ONVIF]device_model VIGI-C540-4G-1.20 [2025-07-25 15:03:07] [ERROR] onvif_start():646 - [ONVIF]g_onvif_port 80 [2025-07-25 15:03:07] [ERROR] onvif_start():649 - [ONVIF]g_onvif_connection_type 0 [2025-07-25 15:03:07] [ERROR] onvif_start():657 - [ONVIF]###### onvif support ###### [2025-07-25 15:03:07] [ERROR] onvif_start():668 - [ONVIF]###### Automatically set static IP ###### [2025-07-25 inet_loop (271): drop_caches: 3 15:03:07] [ERROR] soap_event_create_rb():4644 - [ONVIF]Create http socket 38. [2025-07-25 15:03:07] [ERROR] get_md_capability():4200 - [ONVIF]onvif_mdconf.events is [2025-07-25 15:03:07] [ERROR] get_md_capability():4216 - [ONVIF]md_capability.topic_type is 0 [2025-07-25 15:03:07] [ERROR] onvif_discv_start():172 - [ONVIF]Onvif discovery start [2025-07-25 15:03:07] [ERROR] onvif_srv_start():630 - [ONVIF]Onvif services start, listen_port 2020 [2025-07-25 15:03:07] [ERROR] soap_out_wsdd_hello_type():246 - [ONVIF]g_onvif_port 80 Monitor: ignore <NVMP_START_DONE, 271> because nvmp_no_restart. [2025-07-25 15:03:07] nvmp_print_start_done():317 - [Information][System]Main progress start done error was 0 [2025-07-25 15:03:07] [ERROR] handle_4g_internet_status():9733 - [LTE]4G internet not connected [2025-07-25 15:03:07]lte_start_main_loop():8070 - [LTE]Waiting for usb0 ... [2025-07-25 15:03:07]lte_start_main_loop():8097 - [LTE]Interface usb0 up [2025-07-25 15:03:07] [ERROR] limit_network_speed():1889 - [LTE]cmd:iptables -t filter -I FORWARD 1 -m limit --limit=2000/s --limit-burst=1000 -j ACCEPT; [2025-07-25 15:03:08] [ERROR] limit_network_speed():1893 - [LTE]cmd:iptables -t filter -A FORWARD -j DROP; error was 0 [2025-07-25 15:03:08] [ERROR] readport():451 - [LTE]urc not found, recv [+QIND: PB DONE ] (c540-4g_1.20) login: [2025-07-25 15:03:08] [ERROR] lte_status_check_process():10381 - [LTE]sig_lvl:4 [2025-07-25 15:03:09] [ERROR] get_flow_bytes():4076 - [LTE]read_buf:438,data:438 [2025-07-25 15:03:09] [ERROR] get_flow_bytes():4076 - [LTE]read_buf:300,data:300 [2025-07-25 15:03:09]lte_start_main_loop():8133 - [LTE]Main loop net connecting ... [2025-07-25 15:03:09] [ERROR] handle_4g_internet_status():9733 - [LTE]4G internet not connected error was 0 [2025-07-25 15:03:09] [ERROR] __enlarge_bdesc_ring():596 - [AVDM]avdm0 bdesc ring enlarge 5 to 20 [2025-07-25 15:03:09] [ERROR] lte_status_check_process():10381 - [LTE]sig_lvl:4 [2025-07-25 15:03:09] [ERROR] get_flow_bytes():4076 - [LTE]read_buf:876,data:876 [2025-07-25 15:03:09] [ERROR] get_flow_bytes():4076 - [LTE]read_buf:372,data:372 [2025-07-25 15:03:10] [ERROR] get_runtime_param():1328 - [LTE]start:+CGCONTRDP: 2,5,"ctnet","10.76.50.247","","218.6.200.139","61.139.2.69","","",0,0 +CGCONTRDP: 2,5,"ctnet","254.128.0.0.0.0.0.0.0.0.0.0.0.0.0.1","","36.14.0.86.64.0.128.0.0.0.0.0.0.0.0.105","36.14.0.86.64.0.0.0.0.0.0.0.0.0.2.24","","",0,0 OK . [2025-07-25 15:03:10] [ERROR] get_runtime_param():1402 - [LTE]ipv6:fe80::1 [2025-07-25 15:03:10] [ERROR] get_runtime_param():1412 - [LTE]dnsv6_1:240e:56:4000:8000::69 [2025-07-25 15:03:10] [ERROR] get_runtime_param():1427 - [LTE]dnsv6_2:240e:56:4000::218 [2025-07-25 15:03:10]lte_reload():12064 - [LTE]lte_reload [2025-07-25 15:03:10] [ERROR] ds_pack_config():978 - [DS]encrypt_key_type:0 error was 0 [2025-07-25 15:03:10] [ERROR] get_runtime_mtu():1206 - [LTE]mtu_val:0, old_mtu_val:1480, ret_cnt:1 [2025-07-25 15:03:10] [ERROR] get_runtime_mtu():1216 - [LTE]get lte ipv6 dns config [2025-07-25 15:03:10] [ERROR] get_runtime_mtu():1240 - [LTE]Could not get MTU value by command AT*PCO=2, keep old mtu value. [2025-07-25 15:03:10] [ERROR] lte_start_main_loop():8333 - [LTE]get ipv4 successful, get_ipaddr_flag:17 [2025-07-25 15:03:10] [ERROR] lte_start_main_loop():8338 - [LTE]get ipv6 successful, get_ipaddr_flag:17 [2025-07-25 15:03:10] [ERROR] lte_start_main_loop():8344 - [LTE]info_internet.ipv6_link_status:2 [2025-07-25 15:03:10] [ERROR] ds_pack_config():978 - [DS]encrypt_key_type:0 [2025-07-25 15:03:10] [ERROR] lte_start_main_loop():8378 - [LTE]dhcpc_process_cur_time - dhcpc_process_start_time:0 s [2025-07-25 15:03:10]lte_start_main_loop():8133 - [LTE]Main loop net connecting ... [2025-07-25 15:03:10] [ERROR] onvif_cfg_monitor_reload():111 - [ONVIF]config /protocol/ipv6 changed [2025-07-25 15:03:10] ipv6_enabled():649 - [Information][Network][IPV6]Start IPv6 stateless successful [2025-07-25 15:03:10] ipv6_reload():826 - [Information][Network][IPV6]Enable IPv6 [2025-07-25 15:03:10] [ERROR] nifc_set_if_usb0_info_callback():899 - [NIFC]get usb0 information success [2025-07-25 15:03:10] [ERROR] nifc_set_if_usb0_info_callback():900 - [NIFC]Start set usb0 network information [2025-07-25 15:03:10] [ERROR] nifc_set_if_usb0_info_callback():911 - [NIFC]start set usb0 ipv4 net. [2025-07-25 15:03:10] [ERROR] system_if_del_mac_of_gateway():502 - [NIFC]Ioctl SIOCDARP error, error info:No such device or address [2025-07-25 15:03:10] [ERROR] nifc_set_usb0_net():1579 - [NIFC][LTE] bridge.host_route:0.0.0.0 [2025-07-25 15:03:10] [ERROR] nifc_set_usb0_net():1637 - [NIFC]set usb0 net success [2025-07-25 15:03:10] [ERROR] write_dns():249 - [NIFC]start write ipv6 dns nameserver. [2025-07-25 15:03:11] [ERROR] handle_4g_internet_status():9733 - [LTE]4G internet not connected [2025-07-25 15:03:11] [ERROR] __enlarge_udesc_list():703 - [AVDM]avdm0 udesc list enlarge 200 to 1000 [2025-07-25 15:03:11] md_alarm_start_effect():2769 - [Alarm][Motion Detection]Motion detection start [2025-07-25 15:03:11] [ERROR] lte_status_check_process():10381 - [LTE]sig_lvl:4 [2025-07-25 15:03:11] [ERROR] get_flow_bytes():4076 - [LTE]read_buf:4382,data:4382 [2025-07-25 15:03:11] [ERROR] get_flow_bytes():4076 - [LTE]read_buf:5176,data:5176 [2025-07-25 15:03:11] [ERROR] lte_wan6_stopXlat():5673 - [LTE]lte_wan6_stopXlat begin. [2025-07-25 15:03:11] [ERROR] lte_wan6_stopXlat():5685 - [LTE]virtualIpv4:192.0.0.2 [2025-07-25 15:03:11] [ERROR] lte_wan6_stopXlat():5693 - [LTE]cmd:ip6tables -D FORWARD -t filter -i clatd -j ACCEPT. ip6tables: Bad rule (does a matching rule exist in that chain?). [2025-07-25 15:03:11] [ERROR] lte_wan6_stopXlat():5697 - [LTE]cmd:ip6tables -D FORWARD -t filter -i usb0 -j ACCEPT. ip6tables: Bad rule (does a matching rule exist in that chain?). [2025-07-25 15:03:12] [ERROR] lte_wan6_stopXlat():5701 - [LTE]cmd:iptables -D FORWARD -t filter -i eth0 -j ACCEPT. [2025-07-25 15:03:12] [ERROR] validate_recv_handle():155 - [cloud-service]validate "err_code":-20501, msg:Device id not found. iptables: Bad rule (does a matching rule exist in that chain?). [2025-07-25 15:03:12] [ERROR] lte_wan6_stopXlat():5705 - [LTE]cmd:iptables -D POSTROUTING -t nat -o clatd -j SNAT --to-source 192.0.0.2. iptables: No chain/target/match by that name. [2025-07-25 15:03:12] [ERROR] lte_wan6_stopXlat():5709 - [LTE]cmd:echo "del clatd" > /proc/net/nat46/control. nat46: deleting device (clatd) Could not find device 'clatd' [2025-07-25 15:03:12] [ERROR] lte_wan6_stopXlat():5712 - [LTE]lte_wan6_stopXlat end. [2025-07-25 15:03:12] [ERROR] lte_dhcps_start():687 - [LTE]dev_name:eth0 [2025-07-25 15:03:12] [ERROR] start_ipv4_nat_conversion():1790 - [LTE]cmd:echo '1' > /proc/sys/net/ipv4/ip_forward [2025-07-25 15:03:12] dhcps_start_cb():1044 - [Information][Network][DHCP]Start DHCP server [2025-07-25 15:03:12] [ERROR] start_ipv4_nat_conversion():1826 - [LTE]cmd:iptables -t nat -A PREROUTING -d 192.168.43.1 -p udp -m udp --dport 53 -j DNAT --to-destination 218.6.200.139; [2025-07-25 15:03:12] [ERROR] start_ipv4_nat_conversion():1855 - [LTE]cmd:iptables -t nat -A PREROUTING -d 192.168.43.1 -p tcp -m tcp --dport 53 -j DNAT --to-destination 218.6.200.139; [2025-07-25 15:03:12] [ERROR] start_ipv4_nat_conversion():1863 - [LTE]cmd:iptables -t nat -A POSTROUTING -s 192.168.43.0/255.255.255.0 -o usb0 -j MASQUERADE; [2025-07-25 15:03:12]update_led_state():11558 - [LTE]LED send led event 21 [2025-07-25 15:03:12]lte_start_main_loop():8582 - [LTE]Main loop net check ... [2025-07-25 15:03:12] lte_start_main_loop():8554 - [Information][Network][4G]*** 4g internet connected. [2025-07-25 15:03:12] [ERROR] long_connect_reload():4392 - [LONG_CONNECT]4g internet link_status change to :1 [2025-07-25 15:03:12] [ERROR] remote_log_reload():434 - [REMOTE_LOG]4g internet link_status change to :1 [2025-07-25 15:03:12] [ERROR] ddns_reload():693 - [DDNS]LTE mode: link_status change to :1 [2025-07-25 15:03:12] [ERROR] start_ddns_process():579 - [DDNS]DDNS start [2025-07-25 15:03:12] [ERROR] start_ddns_process():582 - [DDNS]dns is not enable [2025-07-25 15:03:13] md_alarm_end_effect():2837 - [Alarm][Motion Detection]Motion detection end
09-20
<template> <div class="vote-container"> <!-- 投票人信息 --> <div class="voter-info" v-if="voterName && voterIdCard"> <p>投票人:{{ voterName }}</p> <p>身份证:{{ formattedIdCard }}</p> </div> <!-- 投票统计信息 --> <div class="stats"> <div class="stat"> <h3>经理投票</h3> <div class="progress"> <div class="progress-bar" :style="{ width: (votes.A / 5) * 100 + '%' }"></div> </div> <p>{{ votes.A }} / 5</p> </div> <div class="stat"> <h3>厂长投票</h3> <div class="progress"> <div class="progress-bar" :style="{ width: (votes.B / 5) * 100 + '%' }"></div> </div> <p>{{ votes.B }} / 5</p> </div> <div class="stat"> <h3>副厂长投票</h3> <div class="progress"> <div class="progress-bar" :style="{ width: (votes.C / 15) * 100 + '%' }"></div> </div> <p>{{ votes.C }} / 15</p> </div> <div class="stat"> <h3>总票数</h3> <div class="progress"> <div class="progress-bar" :style="{ width: (totalVotes / 25) * 100 + '%' }"></div> </div> <p>{{ totalVotes }} / 25</p> </div> </div> <!-- 被投票人列表 --> <div class="voters-grid"> <div v-for="voter in voters" :key="voter.id" class="voter-card"> <h4>{{ voter.name }}</h4> <p class="voter-id">ID: {{ voter.id }}</p> <div class="vote-options"> <button @click="castVote(voter, 'A')" :disabled="!canVote(voter, 'A')" :class="{ 'selected': voter.vote === 'A', 'disabled': !canVote(voter, 'A') }" > 经理 <!-- 修改为中文 --> </button> <button @click="castVote(voter, 'B')" :disabled="!canVote(voter, 'B')" :class="{ 'selected': voter.vote === 'B', 'disabled': !canVote(voter, 'B') }" > 厂长 <!-- 修改为中文 --> </button> <button @click="castVote(voter, 'C')" :disabled="!canVote(voter, 'C')" :class="{ 'selected': voter.vote === 'C', 'disabled': !canVote(voter, 'C') }" > 副厂长 <!-- 修改为中文 --> </button> </div> </div> </div> <!-- 操作按钮 --> <div class="action-buttons"> <button @click="submitVotes" :disabled="totalVotes === 0 || isSubmitting">提交投票</button> <button @click="resetVotes" :disabled="isSubmitting">重置投票</button> </div> </div> </template> <script setup> import { ref, reactive, computed } from 'vue'; import { useRoute } from 'vue-router'; // 获取路由信息 const route = useRoute(); // 添加用于存储投票人信息的变量 const voterName = ref(''); const voterIdCard = ref(''); // 格式化身份证显示(安全脱敏) const formattedIdCard = computed(() => { if (!voterIdCard.value) return ''; // 显示前6位和后4位,中间用*代替 return voterIdCard.value.substring(0, 6) + '******' + voterIdCard.value.substring(voterIdCard.value.length - 4); }); onMounted(() => { // 从sessionStorage获取数据并立即清除 const voterInfo = sessionStorage.getItem('voterInfo'); if (voterInfo) { const { name, idCard } = JSON.parse(voterInfo); voterName.value = name; voterIdCard.value = idCard; // 关键:立即清除存储防止数据残留 sessionStorage.removeItem('voterInfo'); } }); console.log(routeUserName ) // 生成200个被投票人数据 const voters = ref([]); const generateVoters = () => { const names = []; // 生成随机中文姓名(简单实现) const surnames = ['王', '李', '张', '刘', '陈', '杨', '赵', '黄', '周', '吴']; const names1 = ['伟', '芳', '娜', '洋', '勇', '杰', '明', '霞', '秀', '立']; for (let i = 1; i <= 200; i++) { const surname = surnames[Math.floor(Math.random() * surnames.length)]; const namePart = names1[Math.floor(Math.random() * names1.length)]; voters.value.push({ id: i, name: `${surname}${namePart}`, vote: null // 初始没有投票 }); } }; // 初始化投票人列表 generateVoters(); // 投票统计 const votes = reactive({ A: 0, B: 0, C: 0 }); // 计算总票数 const totalVotes = computed(() => { return votes.A + votes.B + votes.C; }); // 投票方法(优化) const canVote = (voter, type) => { // 情况1:用户取消当前选择的类型(总是允许) if (voter.vote === type) return true; // 情况2:用户从其他类型转换到当前类型 if (voter.vote && voter.vote !== type) { if (type === 'A' && votes.A >= 5) return false; if (type === 'B' && votes.B >= 5) return false; if (type === 'C' && votes.C >= 15) return false; } // 情况3:用户首次投票 if (!voter.vote) { if (type === 'A' && votes.A >= 5) return false; if (type === 'B' && votes.B >= 5) return false; if (type === 'C' && votes.C >= 15) return false; if (totalVotes.value >= 25) return false; } return true; }; // 投票方法(要优化) const castVote = (voter, type) => { // 如果已投票且点击相同类型,取消投票 if (voter.vote === type) { voter.vote = null; votes[type]--; return; } // 如果之前有投票,先取消 if (voter.vote !== null) { votes[voter.vote]--; } // 投新票 voter.vote = type; votes[type]++; }; //投票人信息 // 添加投票人信息数据模型 const voterInfo = reactive({ name: '', idNumber: '' }); // 添加基本信息验证 const isValid = computed(() => { return voterInfo.name.trim() !== '' && voterInfo.idNumber.trim() !== '' && /^\d{17}[\dXx]$/.test(voterInfo.idNumber); }); // 提交投票 // 防止重复提交 const isSubmitting = ref(false); // 提交投票到API const submitVotes = async () => { if (!isValid.value) { alert('请填写正确的姓名和身份证信息'); return; } // 防止重复提交 if (isSubmitting.value) return; isSubmitting.value = true; try { // 构建请求数据 const requestData = { voterInfo: { name: voterInfo.name, idNumber: voterInfo.idNumber }, voteStatistics: { 经理: votes.A, 厂长: votes.B, 副厂长: votes.C }, voteDetails: voters.value .filter(voter => voter.vote !== null) .map(voter => ({ id: voter.id, name: voter.name, voteType: voter.vote === 'A' ? '经理' : voter.vote === 'B' ? '厂长' : '副厂长' })) }; // 发送POST请求 const response = await fetch('/api/wechat/getInvestigate', { method: 'POST', headers: { 'Content-Type': 'application/json' }, body: JSON.stringify(requestData) }); if (!response.ok) { throw new Error(`HTTP error! status: ${response.status}`); } const result = await response.json(); alert('投票提交成功!'); console.log('API响应:', result); // 可选:提交成功后重置表单 resetVotes(); } catch (error) { console.error('提交失败:', error); alert('投票提交失败,请重试'); } finally { isSubmitting.value = false; } }; // 重置投票 const resetVotes = () => { if (confirm('确定要重置所有投票吗?')) { voters.value.forEach(voter => { voter.vote = null; }); votes.A = 0; votes.B = 0; votes.C = 0; voterInfo.name = ''; voterInfo.idNumber = ''; } }; </script> <style scoped> /* 移动端垂直布局 */ @media (max-width: 480px) { .input-group { flex-direction: column; } } /* 平板/桌面端水平布局 */ @media (min-width: 768px) { .input-group { flex-direction: row; } } .vote-container { max-width: 1200px; margin: 0 auto; padding: 20px; } .stats { display: flex; justify-content: space-between; margin-bottom: 30px; background: #f5f7fa; padding: 20px; border-radius: 8px; box-shadow: 0 2px 10px rgba(0, 0, 0, 0.1); } .stat { flex: 1; text-align: center; padding: 0 15px; } .progress { height: 20px; background: #e0e0e0; border-radius: 10px; margin: 10px 0; overflow: hidden; } .progress-bar { height: 100%; background: #3498db; transition: width 0.3s; } .voters-grid { display: grid; grid-template-columns: repeat(auto-fill, minmax(240px, 1fr)); gap: 20px; } .voter-card { background: white; border-radius: 8px; padding: 15px; box-shadow: 0 2px 8px rgba(0, 0, 0, 0.1); transition: transform 0.2s; } .voter-card:hover { transform: translateY(-5px); } .voter-id { color: #777; font-size: 0.9rem; margin-bottom: 15px; } .vote-options { display: flex; justify-content: space-between; } .vote-options button { flex: 1; margin: 0 5px; padding: 8px 0; border: none; border-radius: 4px; cursor: pointer; transition: all 0.2s; } .vote-options button:not(.selected):hover { opacity: 0.9; transform: scale(1.05); } .vote-options button:first-child { background: #ff6b6b; color: white; } .vote-options button:nth-child(2) { background: #4ecdc4; color: white; } .vote-options button:last-child { background: #ffd166; color: white; } .selected { border: 2px solid #2c3e50 !important; font-weight: bold; box-shadow: 0 0 2 rgba(61, 60, 60, 0.5); } .disabled { opacity: 0.5 !important; cursor: not-allowed !important; } .action-buttons { margin-top: 30px; display: flex; justify-content: center; gap: 20px; } .action-buttons button { padding: 12px 30px; border: none; border-radius: 6px; cursor: pointer; font-size: 1rem; font-weight: 600; transition: all 0.2s; } .action-buttons button:first-child { background: #3498db; color: white; } .action-buttons button:first-child:disabled { background: #bdc3c7; cursor: not-allowed; } .action-buttons button:last-child { background: #e74c3c; color: white; } .action-buttons button:hover:not(:disabled) { transform: translateY(-2px); box-shadow: 0 4px 8px rgba(0, 0, 0, 0.2); } </style>报错:Uncaught runtime errors: × ERROR onMounted is not defined ReferenceError: onMounted is not defined at setup (webpack-internal:///./node_modules/babel-loader/lib/index.js??clonedRuleSet-40.use[0]!./node_modules/vue-loader/dist/index.js??ruleSet[0].use[0]!./src/views/Vote.vue?vue&type=script&setup=true&lang=js:41:5) at callWithErrorHandling (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:387:19) at setupStatefulComponent (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:7161:25) at setupComponent (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:7122:36) at mountComponent (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:4835:7) at processComponent (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:4813:9) at patch (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:4488:11) at ReactiveEffect.componentUpdateFn [as fn] (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:5011:9) at ReactiveEffect.run (webpack-internal:///./node_modules/@vue/reactivity/dist/reactivity.esm-bundler.js:323:19) at ReactiveEffect.runIfDirty (webpack-internal:///./node_modules/@vue/reactivity/dist/reactivity.esm-bundler.js:359:12) ERROR Cannot read properties of null (reading 'component') TypeError: Cannot read properties of null (reading 'component') at locateNonHydratedAsyncRoot (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:5592:41) at locateNonHydratedAsyncRoot (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:5597:14) at ReactiveEffect.componentUpdateFn [as fn] (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:4966:40) at ReactiveEffect.run (webpack-internal:///./node_modules/@vue/reactivity/dist/reactivity.esm-bundler.js:323:19) at ReactiveEffect.runIfDirty (webpack-internal:///./node_modules/@vue/reactivity/dist/reactivity.esm-bundler.js:359:12) at callWithErrorHandling (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:387:33) at flushJobs (webpack-internal:///./node_modules/@vue/runtime-core/dist/runtime-core.esm-bundler.js:587:9)
06-08
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值