The operation cannot be completed. See the details.

本文详细介绍了当尝试安装Android Native Development Tools时遇到错误提示'无法完成安装,因为一个或多个必需项未找到'的解决方案。通过两种方法解决资源未下载导致的安装失败问题:一种是复制并运行特定目录的android.bat文件;另一种是下载并替换tools目录中的文件。此外,还提供了重新运行SDKSetup.exe步骤来完成安装。
The operation cannot be completed.  See the details.


Cannot complete the install because one or more required items could not be found.
  Software being installed: Android Native Development Tools 21.1.0.v201302060044-569685 


(com.android.ide.eclipse.ndk.feature.group 21.1.0.v201302060044-569685)
  Missing requirement: Android Native Development Tools 21.1.0.v201302060044-569685 


(com.android.ide.eclipse.ndk.feature.group 21.1.0.v201302060044-569685) requires 


'org.eclipse.cdt.feature.group 0.0.0' but it could not be found



上面的这种错误原因是一些资源没有下载到  但是这些资源SDK.setup或者SDK Manager又无法下载 解决办法如下

解决方法有两种:

第一种方法(老外研究的方法):

1.复制 tools目录 为一个新的目录 tools-copy ,此时在android-sdk-windows 目录下有两个目录 tools 和 tools-copy
2.在tools-copy目录运行 android.bat ,这样就可以正常 update all 了
3.重新运行SDK Setup.exe

 

第二种方法(推荐):

 1.进入sdk的temp文件夹,下载好后会有一个tools_r10-windows.zip(版本号可能会不一样,如r09)
 2.解压此文件,将解压出来的文件覆盖tools中的文件
 3.重新运行SDK Setup.exe


老公 你帮我看一下 我是不是可以去睡觉 等它自己编译就好了?”PowerShell 7 环境已加载 (版本: 7.5.2) PowerShell 7 环境已加载 (版本: 7.5.2) PS C:\Users\Administrator\Desktop> cd E:\PyTorch_Build\pytorch PS E:\PyTorch_Build\pytorch> python -m venv rtx5070_env PS E:\PyTorch_Build\pytorch> .\rtx5070_env\Scripts\activate (rtx5070_env) PS E:\PyTorch_Build\pytorch> # 1. 运行修复脚本 (rtx5070_env) PS E:\PyTorch_Build\pytorch> .\complete_install_fix.ps1 尝试从备用源下载: https://pilotfiber.dl.sourceforge.net/project/openblas/v0.3.27/OpenBLAS-0.3.27-x64.zip ❌ 下载失败: The variable '$localZip' cannot be retrieved because it has not been set. 尝试从备用源下载: https://cfhcable.dl.sourceforge.net/project/openblas/v0.3.27/OpenBLAS-0.3.27-x64.zip ❌ 下载失败: The variable '$localZip' cannot be retrieved because it has not been set. 尝试从备用源下载: https://superb-sea2.dl.sourceforge.net/project/openblas/v0.3.27/OpenBLAS-0.3.27-x64.zip ❌ 下载失败: The variable '$localZip' cannot be retrieved because it has not been set. 尝试从GitHub直接下载... InvalidOperation: E:\PyTorch_Build\pytorch\complete_install_fix.ps1:25 Line | 25 | Invoke-WebRequest -Uri $githubRaw -OutFile $localZip | ~~~~~~~~~ | The variable '$localZip' cannot be retrieved because it has not been set. InvalidOperation: E:\PyTorch_Build\pytorch\complete_install_fix.ps1:35 Line | 35 | [System.IO.Compression.ZipFile]::ExtractToDirectory($localZip, $newIn … | ~~~~~~~~~ | The variable '$localZip' cannot be retrieved because it has not been set. Find-BLASBinary: E:\PyTorch_Build\pytorch\complete_install_fix.ps1:38 Line | 38 | $blasFile = Find-BLASBinary -SearchPath $newInstallPath | ~~~~~~~~~~~~~~~ | The term 'Find-BLASBinary' is not recognized as a name of a cmdlet, function, script file, or executable | program. Check the spelling of the name, or if a path was included, verify that the path is correct and try | again. ⚠️ 二进制分发版无效,尝试从源码编译... Chocolatey v2.5.0 3 validations performed. 2 success(es), 1 warning(s), and 0 error(s). Validation Warnings: - A pending system reboot request has been detected, however, this is being ignored due to the current Chocolatey configuration. If you want to halt when this occurs, then either set the global feature using: choco feature enable --name="exitOnRebootDetected" or pass the option --exit-when-reboot-detected. Installing the following packages: cmake By installing, you accept licenses for the packages. Downloading package from source 'https://community.chocolatey.org/api/v2/' Progress: Downloading cmake.install 4.1.1... 100% cmake.install v4.1.1 [Approved] cmake.install package files install completed. Performing other installation steps. Installing 64-bit cmake.install... cmake.install has been installed. The install of cmake.install was successful. Software installed as 'msi', install location is likely default. Downloading package from source 'https://community.chocolatey.org/api/v2/' Progress: Downloading cmake 4.1.1... 100% cmake v4.1.1 [Approved] cmake package files install completed. Performing other installation steps. The install of cmake was successful. Deployed to 'C:\ProgramData\chocolatey\lib\cmake' Chocolatey installed 2/2 packages. See the log for details (C:\ProgramData\chocolatey\logs\chocolatey.log). Chocolatey v2.5.0 3 validations performed. 2 success(es), 1 warning(s), and 0 error(s). Validation Warnings: - A pending system reboot request has been detected, however, this is being ignored due to the current Chocolatey configuration. If you want to halt when this occurs, then either set the global feature using: choco feature enable --name="exitOnRebootDetected" or pass the option --exit-when-reboot-detected. Installing the following packages: mingw By installing, you accept licenses for the packages. Downloading package from source 'https://community.chocolatey.org/api/v2/' Progress: Downloading mingw 13.2.0... 100% mingw v13.2.0 [Approved] mingw package files install completed. Performing other installation steps. Downloading mingw 64 bit from 'https://github.com/niXman/mingw-builds-binaries/releases/download/13.2.0-rt_v11-rev0/x86_64-13.2.0-release-posix-seh-ucrt-rt_v11-rev0.7z' Progress: 73% - Saving 51.17 MB of 69.57 MB“
09-03
Your feedback is requested on changes under consideration for SLF4J version 2.1.0. SLF4J warning or error messages and their meanings No SLF4J providers were found. This message is a warning and not an error. It is reported when no SLF4J providers could be found on the class path. SLF4J requires a logging provider because it is a logging API and not an logging implementation. Placing one (and only one) of the many available providers such as slf4j-nop.jar slf4j-simple.jar, slf4j-reload4j.jar, slf4j-jdk14.jar or logback-classic.jar on the class path will solve the problem. In the absence of a provider, SLF4J will default to a no-operation (NOP) logger provider. Please note that slf4j-api version 2.0.x and later use the ServiceLoader mechanism. Earlier versions relied on the static binder mechanism which is no longer honored by slf4j-api. Please read the FAQ entry What has changed in SLF4J version 2.0.0? for further important details. If you are responsible for packaging an application and do not care about logging, then placing slf4j-nop.jar on the class path of your application will get rid of this warning message. Note that embedded components such as libraries or frameworks should not declare a dependency on any SLF4J providers but only depend on slf4j-api. When a library declares a compile-time dependency on a SLF4J provider, it imposes that provider on the end-user, thus negating SLF4J's purpose. Class path contains SLF4J bindings targeting slf4j-api versions 1.7.x or earlier Planning for the advent of Jigsaw (Java 9), slf4j-api version 2.0.x and later use the ServiceLoader mechanism. Earlier versions of SLF4J relied on the static binder mechanism which is no longer honored by slf4j-api version 2.0.x. In case SLF4J 2.x finds no providers targeting SLF4J 2.x but finds instead bindings targeting SLF4J 1.7 or earlier, it will list the bindings it finds but otherwise will ignore them. This can be solved by placing an SLF4J provider on your classpath, such providers include logback version 1.3.x and later, as well as one of slf4j-reload4j, slf4j-jdk14, slf4j-simple version 2.0.0 or later. See also the FAQ entry What has changed in SLF4J version 2.0.0? for further important details. IllegalStateException: org.slf4j.LoggerFactory in failed state. Original exception was thrown EARLIER. This IllegalStateException is thrown post-initialization and informs the user that initialization of LoggerFactory has failed. Note It is important to realize that the exception causing the failure was thrown at an earlier stage. This earlier exception should provide more valuable information about the root cause of the problem. The method o.a.commons.logging.impl.SLF4FLogFactory#release was invoked. Given the structure of the commons-logging API, in particular as implemented by SLF4J, the o.a.commons.logging.impl.SLF4FLogFactory#release() method should never be called. However, depending on the deployment of commons-logging.jar files in your servlet container, release() method may be unexpectedly invoked by a copy of org.apache.commons.logging.LogFactory class shipping with commons-logging.jar. This is a relatively common occurrence with recent versions of Tomcat, especially if you place jcl-over-slf4j.jar in WEB-INF/lib directory of your web-application instead of $TOMCAT_HOME/common/lib, where $TOMCAT_HOME stands for the directory where Tomcat is installed. In order to fully benefit from the stability offered by jcl-over-slf4j.jar, we recommend that you place jcl-over-slf4j.jar in $TOMCAT_HOME/common/lib without placing a copy in your web-applications. Please also see bug #22. Operation [suchAndSuch] is not supported in jcl-over-slf4j. An UnsupportedOperationException is thrown whenever one of the protected methods introduced in JCL 1.1 are invoked. These methods are invoked by LogFactory implementations shipping with commons-logging.jar. However, the LogFactory implemented by jcl-over-slf4j.jar, namely SLF4FLogFactory, does not call any of these methods. If you observe this problem, then it is highly probable that you have a copy of commons-logging.jar in your class path overriding the classes shipping with jcl-over-slf4j.jar. Note that this issue is very similar in nature to the warning issued when the "o.a.commons.logging.impl.SLF4FLogFactory.release()" method is invoked, discussed in the previous item. Detected logger name mismatch Logger name mismatch warnings are printed only if the slf4j.detectLoggerNameMismatch system property is set to true. By default, this property is not set and no warnings will be printed even in case of a logger name mismatch. since 1.7.9 The warning will be printed in case the name of the logger specified via a class passed as an argument to the LoggerFactory.getLogger(Class) method differs from the name of the caller as computed internally by SLF4J. For example, the following code snippet package com.acme; import com.foo.Kangaroo; class Fruit { Logger logger = LoggerFactory.getLogger(Kangaroo.class); } will result in the warning SLF4J: Detected logger name mismatch. Given name: "com.foo.Kangaroo"; computed name: "com.acme.Fruit". but only if slf4j.detectLoggerNameMismatch system property is set to true. No warning will be issued for the special case where the class in which the logger is defined is a super-type of the class parameter passed as argument. For example, class A { Logger logger = LoggerFactory.getLogger(getClass()); } class B extends A { // no mismatch warning will be issued when B is instantiated // given that class A is a super-type of class B } If you come across a mismatch warning which cannot be explained, then you might have spotted a white elephant, that is a very rare occurrence where SLF4J cannot correctly compute the name of the class where a logger is defined. We are very interested to learn about such cases. If and when you spot an inexplicable mismatch, please do file a bug report with us. Failed to load class org.slf4j.impl.StaticLoggerBinder This warning message is reported by slf4j-api version 1.7.x and earlier when the org.slf4j.impl.StaticLoggerBinder class could not be loaded into memory. This happens when no appropriate SLF4J binding could be found on the class path. Placing one (and only one) of slf4j-nop.jar slf4j-simple.jar, slf4j-log4j12.jar, slf4j-jdk14.jar or logback-classic.jar on the class path should solve the problem. If you are seeing this message, then you are NOT using slf4j-api version 2.0 or later but slf4j-api version 1.7.x or earlier. Slf4j-api versions 2.0.x and later use the ServiceLoader mechanism. Backends such as logback 1.3 and later which target slf4j-api 2.x, do not ship with org.slf4j.impl.StaticLoggerBinder. If you place a logging backend which targets slf4j-api 2.0.x, you need slf4j-api-2.x.jar on the classpath. See also relevant faq entry. since 1.6.0 As of SLF4J version 1.6, in the absence of a binding, SLF4J will default to a no-operation (NOP) logger implementation. If you are responsible for packaging an application and do not care about logging, then placing slf4j-nop.jar on the class path of your application will get rid of this warning message. Note that embedded components such as libraries or frameworks should not declare a dependency on any SLF4J binding (or provider) but only depend on slf4j-api. When a library declares a compile-time dependency on a SLF4J binding (or provider), it imposes that binding (or provider) on the end-user, thus negating SLF4J's purpose. Multiple bindings were found on the class path SLF4J API is designed to bind with one and only one underlying logging framework at a time. If more than one binding is present on the class path, SLF4J will emit a warning, listing the location of those bindings. When multiple bindings are available on the class path, select one and only one binding you wish to use, and remove the other bindings. For example, if you have both slf4j-simple-2.0.17.jar and slf4j-nop-2.0.17.jar on the class path and you wish to use the nop (no-operation) binding, then remove slf4j-simple-2.0.17.jar from the class path. The list of locations that SLF4J provides in this warning usually provides sufficient information to identify the dependency transitively pulling in an unwanted SLF4J binding into your project. In your project's pom.xml file, exclude this SLF4J binding when declaring the unscrupulous dependency. For example, cassandra-all version 0.8.1 declares both log4j and slf4j-log4j12 as compile-time dependencies. Thus, when you include cassandra-all as a dependency in your project, the cassandra-all declaration will cause both slf4j-log4j12.jar and log4j.jar to be pulled in as dependencies. In case you do not wish to use log4j as the SLF4J backend, you can instruct Maven to exclude these two artifacts as shown next: <dependencies> <dependency> <groupId> org.apache.cassandra</groupId> <artifactId>cassandra-all</artifactId> <version>0.8.1</version> <exclusions> <exclusion> <groupId>org.slf4j</groupId> <artifactId>slf4j-log4j12</artifactId> </exclusion> <exclusion> <groupId>log4j</groupId> <artifactId>log4j</artifactId> </exclusion> </exclusions> </dependency> </dependencies> Note The warning emitted by SLF4J is just that, a warning. Even when multiple bindings are present, SLF4J will pick one logging framework/implementation and bind with it. The way SLF4J picks a binding is determined by the JVM and for all practical purposes should be considered random. As of version 1.6.6, SLF4J will name the framework/implementation class it is actually bound to. Embedded components such as libraries or frameworks should not declare a dependency on any SLF4J binding but only depend on slf4j-api. When a library declares a compile-time dependency on a SLF4J binding, it imposes that binding on the end-user, thus negating SLF4J's purpose. When you come across an embedded component declaring a compile-time dependency on any SLF4J binding, please take the time to contact the authors of said component/library and kindly ask them to mend their ways. slf4j-api version does not match that of the binding An SLF4J binding designates an artifact such as slf4j-jdk14.jar or slf4j-log4j12.jar used to bind slf4j to an underlying logging framework, say, java.util.logging and respectively log4j. Mixing different versions of slf4j-api.jar and SLF4J binding (a.k.a. provider since 2.0.0) can cause problems. For example, if you are using slf4j-api-2.0.17.jar, then you should also use slf4j-simple-2.0.17.jar, using slf4j-simple-1.5.5.jar will not work. Note From the client's perspective all versions of slf4j-api are compatible. Client code compiled with slf4j-api-N.jar will run perfectly fine with slf4j-api-M.jar for any N and M. You only need to ensure that the version of your binding matches that of the slf4j-api.jar. You do not have to worry about the version of slf4j-api.jar used by a given dependency in your project. You can always use any version of slf4j-api.jar, and as long as the version of slf4j-api.jar and its binding match, you should be fine. At initialization time, if SLF4J suspects that there may be an api vs. binding version mismatch problem, it will emit a warning about the suspected mismatch. Logging factory implementation cannot be null This error is reported when the LoggerFactory class could not find an appropriate binding. Placing one (and only one) of slf4j-nop.jar, slf4j-simple.jar, slf4j-log4j12.jar, slf4j-jdk14.jar or logback-classic.jar on the class path should prove to be an effective remedy. Detected both log4j-over-slf4j.jar AND slf4j-reload4j on the class path, preempting StackOverflowError. The purpose of slf4j-reload4j module is to delegate or redirect calls made to an SLF4J logger to log4j/reload4j. The purpose of the log4j-over-slf4j module is to redirect calls made to a log4j logger to SLF4J. If SLF4J is bound withslf4j-reload4j.jar and log4j-over-slf4j.jar is also present on the class path, a StackOverflowError will inevitably occur immediately after the first invocation of an SLF4J or a log4j logger. Here is how the exception might look like: Exception in thread "main" java.lang.StackOverflowError at java.util.Hashtable.containsKey(Hashtable.java:306) at org.apache.log4j.Log4jLoggerFactory.getLogger(Log4jLoggerFactory.java:36) at org.apache.log4j.LogManager.getLogger(LogManager.java:39) at org.slf4j.impl.Log4jLoggerFactory.getLogger(Log4jLoggerFactory.java:73) at org.slf4j.LoggerFactory.getLogger(LoggerFactory.java:249) at org.apache.log4j.Category.<init>(Category.java:53) at org.apache.log4j.Logger..<init>(Logger.java:35) at org.apache.log4j.Log4jLoggerFactory.getLogger(Log4jLoggerFactory.java:39) at org.apache.log4j.LogManager.getLogger(LogManager.java:39) at org.slf4j.impl.Log4jLoggerFactory.getLogger(Log4jLoggerFactory.java:73) at org.slf4j.LoggerFactory.getLogger(LoggerFactory.java:249) at org.apache.log4j.Category..<init>(Category.java:53) at org.apache.log4j.Logger..<init>(Logger.java:35) at org.apache.log4j.Log4jLoggerFactory.getLogger(Log4jLoggerFactory.java:39) at org.apache.log4j.LogManager.getLogger(LogManager.java:39) subsequent lines omitted... Since 1.5.11 SLF4J software preempts the inevitable stack overflow error by throwing an exception with details about the actual cause of the problem. This is deemed to be better than leaving the user wondering about the reasons of the StackOverflowError. Note that since reload4j provides the log4j 1.x API, reload4j.jar and log4j-over-slf4j.jar cannot be present simultaneously on your classpath. For more background on this topic see Bridging legacy APIs. Detected both jcl-over-slf4j.jar AND slf4j-jcl.jar on the class path, preempting StackOverflowError. The purpose of slf4j-jcl module is to delegate or redirect calls made to an SLF4J logger to jakarta commons logging (JCL). The purpose of the jcl-over-slf4j module is to redirect calls made to a JCL logger to SLF4J. If SLF4J is bound with slf4j-jcl.jar and jcl-over-slf4j.jar is also present on the class path, then a StackOverflowError will inevitably occur immediately after the first invocation of an SLF4J or a JCL logger. Here is how the exception might look like: Exception in thread "main" java.lang.StackOverflowError at java.lang.String.hashCode(String.java:1482) at java.util.HashMap.get(HashMap.java:300) at org.slf4j.impl.JCLLoggerFactory.getLogger(JCLLoggerFactory.java:67) at org.slf4j.LoggerFactory.getLogger(LoggerFactory.java:249) at org.apache.commons.logging.impl.SLF4JLogFactory.getInstance(SLF4JLogFactory.java:155) at org.apache.commons.logging.LogFactory.getLog(LogFactory.java:289) at org.slf4j.impl.JCLLoggerFactory.getLogger(JCLLoggerFactory.java:69) at org.slf4j.LoggerFactory.getLogger(LoggerFactory.java:249) at org.apache.commons.logging.impl.SLF4JLogFactory.getInstance(SLF4JLogFactory.java:155) subsequent lines omitted... Since 1.5.11 SLF4J software preempts the inevitable stack overflow error by throwing an exception with details about the actual cause of the problem. This is deemed to be better than leaving the user wondering about the reasons of the StackOverflowError. For more background on this topic see Bridging legacy APIs. Exception in thread "main" java.lang.NoSuchFieldError: tlm Exception in thread "main" java.lang.NoSuchFieldError: tlm at org.apache.log4j.MDCFriend.fixForJava9(MDCFriend.java:11) at org.slf4j.impl.Log4jMDCAdapter.(Log4jMDCAdapter.java:38) at Main.main(Main.java:5) The NoSuchFieldError is thrown when slf4j-log4j12 attempts to access the 'tlm' package private field in org.apache.log4j.MDC which was specified as being of type java.lang.Object in log4j 1.2.x but was changed to java.lang.ThreadLocal in reload4j. Moreover, such access to package private fields from different modules is not authorized by default in modularized applications in Java 9 and later. To keep a long story short, the NoSuchFieldError can be avoided by using slf4j-reload4j.jar with reload4j.jar. Stated differently, org.slf4j.MDC cannot be used with the slf4j-log4j12.jar and reload4j.jar combination. Update: The issue described above was fixed in reload4j 1.2.21. Although it is still recommended that you use slf4j-reload4j as the preferred adapter for the slf4j/reload4j combination, with reload4j version 1.2.21 and later you can freely mix any version of slf4j-log4j12, if you need to. Failed to load class "org.slf4j.impl.StaticMDCBinder" This error indicates that appropriate SLF4J binding could not be found on the class path. Placing one (and only one) of slf4j-nop.jar, slf4j-simple.jar, slf4j-log4j12.jar, slf4j-jdk14.jar or logback-classic.jar on the class path should solve the problem. MDCAdapter cannot be null This error is reported when org.slf4j.MDC class has not been initialized correctly. Same cause and remedy as the previously listed item. A number (N) of logging calls during the initialization phase have been intercepted and are now being replayed. These are subject to the filtering rules of the underlying logging system. since 1.7.15 Logging calls made during the initialization phase are recorded and replayed post-initialization. Note that the replayed logging calls are subject to filtering by the underlying logging system. In principle, replaying only occurs for applications which are already multithreaded at the time the first logging call occurs. See also substitute loggers. Substitute loggers were created during the default configuration phase of the underlying logging system Highly configurable logging systems such as logback and log4j may create components which invoke loggers during their own initialization. See issue LOGBACK-127 for a typical occurrence. However, since the binding process with SLF4J has not yet completed (because the underlying logging system was not yet completely loaded into memory), it is not possible to honor such logger creation requests. To avoid this chicken-and-egg problem, SLF4J creates substitute loggers during this phase (initialization). Calls made to the substitute loggers during this phase are simply dropped. After the initialization completes, the substitute logger will delegate logging calls to the appropriate logger implementation and otherwise will function as any other logger returned by LoggerFactory. If any substitute logger had to be created, SLF4J will emit a listing of such loggers. This list is intended to let you know that any logging calls made to these loggers during initialization have been dropped. See also intercepted and replayed logging calls. SLF4J versions 1.4.0 and later requires log4j 1.2.12 or later The trace level was added to log4j in version 1.2.12 released on August 29, 2005. The trace level was added to the SLF4J API in version 1.4.0 on May 16th, 2007. Thus, starting with SLF4J 1.4.0, the log4j binding for SLF4J requires log4j version 1.2.12 or above. However, as reported in issue 59, in some environments it may be difficult to upgrade the log4j version. To accommodate such circumstances, SLF4J's Log4jLoggerAdapter will map the TRACE level as DEBUG. java.lang.NoClassDefFoundError: org/slf4j/event/LoggingEvent Logback-classic version 1.1.4 and later require slf4j-api version 1.7.15 or later. With an earlier slf4j-api.jar in the classpath, attempting introspection of a Logger instance returned by logback version 1.1.4 or later will result in a NoClassDefFoundError similar to that shown below. Exception in thread "main" java.lang.NoClassDefFoundError: org/slf4j/event/LoggingEvent at java.lang.Class.getDeclaredMethods0(Native Method) at java.lang.Class.privateGetDeclaredMethods(Class.java:2451) at java.lang.Class.privateGetPublicMethods(Class.java:2571) at java.lang.Class.getMethods(Class.java:1429) at java.beans.Introspector.getPublicDeclaredMethods(Introspector.java:1261) at java.beans.Introspector.getTargetMethodInfo(Introspector.java:1122) at java.beans.Introspector.getBeanInfo(Introspector.java:414) at java.beans.Introspector.getBeanInfo(Introspector.java:161) Placing slf4j-api.jar version 1.7.15 or later in the classpath should solve the issue. Note that this problem only occurs with logback version 1.1.4 and later, other bindings such as slf4j-log4j, slf4j-jdk14 and slf4j-simple are unaffected.
最新发布
09-24
Linking files for Subswath 1 ... Linking files for Subswath 2 ... Linking files for Subswath 3 ... [1] 22215 [2] 22216 [3] 22217 [3] - Done p2p_processing.csh S1_TOPS ... [2] - Done p2p_processing.csh S1_TOPS ... [1] + Done p2p_processing.csh S1_TOPS ... *** buffer overflow detected ***: terminated Aborted (core dumped) *** buffer overflow detected ***: terminated Aborted (core dumped) *** buffer overflow detected ***: terminated Aborted (core dumped) Merging START Number of Files to be merged is 3 merge_swath [ERROR]: Cannot find file ../F1/intf/2025056_2025104/phasefilt.grd merge_swath [ERROR]: Cannot find file ../F1/intf/2025056_2025104/phasefilt.grd merge_swath [ERROR]: File ../F1/intf/2025056_2025104/phasefilt.grd not found [Session merge_swath (0)]: Error returned from GMT API: GMT_FILE_NOT_FOUND (16) cannot open grids ../F1/intf/2025056_2025104/phasefilt.grd Number of Files to be merged is 3 merge_swath [ERROR]: Cannot find file ../F1/intf/2025056_2025104/corr.grd merge_swath [ERROR]: Cannot find file ../F1/intf/2025056_2025104/corr.grd merge_swath [ERROR]: File ../F1/intf/2025056_2025104/corr.grd not found [Session merge_swath (0)]: Error returned from GMT API: GMT_FILE_NOT_FOUND (16) cannot open grids ../F1/intf/2025056_2025104/corr.grd Number of Files to be merged is 3 merge_swath [ERROR]: Cannot find file ../F1/intf/2025056_2025104/mask.grd merge_swath [ERROR]: Cannot find file ../F1/intf/2025056_2025104/mask.grd merge_swath [ERROR]: File ../F1/intf/2025056_2025104/mask.grd not found [Session merge_swath (0)]: Error returned from GMT API: GMT_FILE_NOT_FOUND (16) cannot open grids ../F1/intf/2025056_2025104/mask.grd Merging END Recomputing the projection LUT... grep: S1_20250226_113936_F1.PRM: 没有那个文件或目录 error - could not open file S1_20250226_113936_F1.PRM couldn't open master.PRM Usage: SAT_llt2rat master.PRM prec [-bo[s|d]] < inputfile > outputfile master.PRM - parameter file for master image and points to LED orbit file precise - (0) standard back geocoding, (1) - polynomial refinenent (slower) inputfile - lon, lat, elevation [ASCII] outputfile - range, azimuth, elevation(ref to radius in PRM), lon, lat [ASCII default] -bos or -bod - binary single or double precision output (only output results within data coverage, PRM num_lines, num_rng_bins ) example: SAT_llt2rat master.PRM 0 < topo.llt > topo.ratll error - could not open file S1_20250226_113936_F1.PRM grdinfo [ERROR]: Cannot find file phasefilt.grd grdinfo [ERROR]: Must specify one or more input files Usage: landmask.csh region_cut[0/10600/0/27648] make a landmask in radar coordinates, needs to run with trans.dat NOTE: The region_cut can be specified in batch.config file SNAPHU.CSH - START threshold_snaphu: 0.1 grdmath [ERROR]: corr_patch.grd is not a number, operator or file name grdmath [ERROR]: corr_patch.grd is not a number, operator or file name grdmath [ERROR]: mask2_patch.grd is not a number, operator or file name grd2xyz [ERROR]: Cannot find file phase_patch.grd grd2xyz [ERROR]: Cannot find file phase_patch.grd grd2xyz [ERROR]: File phase_patch.grd not found [Session gmt (0)]: Error returned from GMT API: GMT_FILE_NOT_FOUND (16) grd2xyz [ERROR]: Cannot find file corr_tmp.grd grd2xyz [ERROR]: Cannot find file corr_tmp.grd grd2xyz [ERROR]: File corr_tmp.grd not found [Session gmt (0)]: Error returned from GMT API: GMT_FILE_NOT_FOUND (16) unwrapping phase with snaphu - higher threshold for faster unwrapping grdinfo [ERROR]: Cannot find file phase_patch.grd grdinfo [ERROR]: Must specify one or more input files snaphu v2.0.4 32 parameters input from file /usr/local/GMTSAR/share/gmtsar/snaphu/config/snaphu.conf.brief (297 lines total) not enough input arguments. type snaphu -h for help grdinfo [ERROR]: Cannot find file phase_patch.grd grdinfo [ERROR]: Must specify one or more input files grdinfo [ERROR]: Cannot find file phase_patch.grd grdinfo [ERROR]: Must specify one or more input files xyz2grd [ERROR]: Must specify -R option xyz2grd [ERROR]: Option -I: Must specify positive increment(s) grdinfo [ERROR]: Cannot find file phase_patch.grd grdinfo [ERROR]: Must specify one or more input files grdinfo [ERROR]: Cannot find file phase_patch.grd grdinfo [ERROR]: Must specify one or more input files xyz2grd [ERROR]: Must specify -R option xyz2grd [ERROR]: Option -I: Must specify positive increment(s) grdmath [ERROR]: tmp.grd is not a number, operator or file name mv: 对 'tmp.grd' 调用 stat 失败: 没有那个文件或目录 grdgradient [ERROR]: Cannot find file unwrap.grd grdinfo [ERROR]: Cannot find file unwrap.grd grdinfo [ERROR]: Must specify one or more input files makecpt [ERROR]: Option T: Must specify valid min[/max[/inc[<unit>|+n]]] option grdimage [ERROR]: Cannot find file unwrap.grd grdimage [ERROR]: Option -I: Must specify intensity file, value, or modifiers psscale [WARNING]: No country code matching unwrap.grd (skipped) psscale [ERROR]: Color palette table unwrap.cpt is empty [Session gmt (0)]: Error returned from GMT API: GMT_CPT_READ_ERROR (8) [Session gmt (0)]: Error returned from GMT API: GMT_CPT_READ_ERROR (8) psconvert [ERROR]: The file unwrap.ps has no BoundingBox in the first 20 lines or last 256 bytes. Use -A option. Unwrapped phase map: unwrap.pdf SNAPHU.CSH - END GEOCODE-START proj_ra2ll.csh grd2xyz [ERROR]: Cannot find file phasefilt.grd grd2xyz [ERROR]: Cannot find file phasefilt.grd grd2xyz [ERROR]: File phasefilt.grd not found [Session gmt (0)]: Error returned from GMT API: GMT_FILE_NOT_FOUND (16) gmtinfo [ERROR]: No input data found! surface [ERROR]: Must specify -R option surface [INFORMATION]: Provides 5, expects 3-column binary data surface [ERROR]: Must specify -R option surface [INFORMATION]: Provides 5, expects 3-column binary data grdtrack [ERROR]: Cannot find file raln.grd grdtrack [ERROR]: Cannot find file ralt.grd grdtrack [ERROR]: Must specify -G at least once gmtconvert [WARNING]: File <stdin> is empty! gmtconvert [WARNING]: No data records provided Sampling in geocoordinates with 50 meter pixels ... gmtinfo [ERROR]: No input data found! range: Subscript out of range. gmtmath [ERROR]: Operation "DIV" requires 2 operands gmtmath [ERROR]: Operation "MUL" requires 2 operands gmtinfo [ERROR]: No input data found! blockmedian [ERROR]: Unable to decode as a floating point number blockmedian [ERROR]: Option -I parsing failure. Correct syntax: -I<xinc>[+e|n][/<yinc>[+e|n]] Specify increment(s) and optionally append units or modifiers. For geographic regions in degrees you can optionally append units from this list: (d)egree [Default], (m)inute, (s)econd, m(e)ter, (f)oot, (k)ilometer, (M)ile, (n)autical mile, s(u)rvey foot. +e Adjust the region to fit increments [Adjust increment to fit domain]. +n Increment specifies the number of nodes instead. Then, the actual increments are calculated from the given domain and node- registration settings (see Appendix B for details). Note: If -R<grdfile> was used then -I (and -R and maybe -r) have been set; use -I to override those increments. blockmedian [ERROR]: Must specify -R option blockmedian [ERROR]: Option -I: Must specify positive increment(s) blockmedian [INFORMATION]: Provides 3, expects 3-column binary data xyz2grd [ERROR]: Unable to decode as a floating point number xyz2grd [ERROR]: Option -I parsing failure. Correct syntax: -I<xinc>[+e|n][/<yinc>[+e|n]] Specify increment(s) and optionally append units or modifiers. For geographic regions in degrees you can optionally append units from this list: (d)egree [Default], (m)inute, (s)econd, m(e)ter, (f)oot, (k)ilometer, (M)ile, (n)autical mile, s(u)rvey foot. +e Adjust the region to fit increments [Adjust increment to fit domain]. +n Increment specifies the number of nodes instead. Then, the actual increments are calculated from the given domain and node- registration settings (see Appendix B for details). Note: If -R<grdfile> was used then -I (and -R and maybe -r) have been set; use -I to override those increments. xyz2grd [ERROR]: Must specify -R option xyz2grd [ERROR]: Option -I: Must specify positive increment(s) grdmath [ERROR]: corr.grd is not a number, operator or file name grdmath [ERROR]: phasefilt.grd is not a number, operator or file name proj_ra2ll.csh grd2xyz [ERROR]: Cannot find file phasefilt_mask.grd grd2xyz [ERROR]: Cannot find file phasefilt_mask.grd grd2xyz [ERROR]: File phasefilt_mask.grd not found [Session gmt (0)]: Error returned from GMT API: GMT_FILE_NOT_FOUND (16) gmtinfo [ERROR]: No input data found! surface [ERROR]: Must specify -R option surface [INFORMATION]: Provides 5, expects 3-column binary data surface [ERROR]: Must specify -R option surface [INFORMATION]: Provides 5, expects 3-column binary data grdtrack [ERROR]: Cannot find file raln.grd grdtrack [ERROR]: Cannot find file ralt.grd grdtrack [ERROR]: Must specify -G at least once gmtconvert [WARNING]: File <stdin> is empty! gmtconvert [WARNING]: No data records provided Sampling in geocoordinates with 50 meter pixels ... gmtinfo [ERROR]: No input data found! range: Subscript out of range. gmtmath [ERROR]: Operation "DIV" requires 2 operands gmtmath [ERROR]: Operation "MUL" requires 2 operands gmtinfo [ERROR]: No input data found! blockmedian [ERROR]: Unable to decode as a floating point number blockmedian [ERROR]: Option -I parsing failure. Correct syntax: -I<xinc>[+e|n][/<yinc>[+e|n]] Specify increment(s) and optionally append units or modifiers. For geographic regions in degrees you can optionally append units from this list: (d)egree [Default], (m)inute, (s)econd, m(e)ter, (f)oot, (k)ilometer, (M)ile, (n)autical mile, s(u)rvey foot. +e Adjust the region to fit increments [Adjust increment to fit domain]. +n Increment specifies the number of nodes instead. Then, the actual increments are calculated from the given domain and node- registration settings (see Appendix B for details). Note: If -R<grdfile> was used then -I (and -R and maybe -r) have been set; use -I to override those increments. blockmedian [ERROR]: Must specify -R option blockmedian [ERROR]: Option -I: Must specify positive increment(s) blockmedian [INFORMATION]: Provides 3, expects 3-column binary data xyz2grd [ERROR]: Unable to decode as a floating point number xyz2grd [ERROR]: Option -I parsing failure. Correct syntax: -I<xinc>[+e|n][/<yinc>[+e|n]] Specify increment(s) and optionally append units or modifiers. For geographic regions in degrees you can optionally append units from this list: (d)egree [Default], (m)inute, (s)econd, m(e)ter, (f)oot, (k)ilometer, (M)ile, (n)autical mile, s(u)rvey foot. +e Adjust the region to fit increments [Adjust increment to fit domain]. +n Increment specifies the number of nodes instead. Then, the actual increments are calculated from the given domain and node- registration settings (see Appendix B for details). Note: If -R<grdfile> was used then -I (and -R and maybe -r) have been set; use -I to override those increments. xyz2grd [ERROR]: Must specify -R option xyz2grd [ERROR]: Option -I: Must specify positive increment(s) proj_ra2ll.csh grd2xyz [ERROR]: Cannot find file corr.grd grd2xyz [ERROR]: Cannot find file corr.grd grd2xyz [ERROR]: File corr.grd not found [Session gmt (0)]: Error returned from GMT API: GMT_FILE_NOT_FOUND (16) gmtinfo [ERROR]: No input data found! surface [ERROR]: Must specify -R option surface [INFORMATION]: Provides 5, expects 3-column binary data surface [ERROR]: Must specify -R option surface [INFORMATION]: Provides 5, expects 3-column binary data grdtrack [ERROR]: Cannot find file raln.grd grdtrack [ERROR]: Cannot find file ralt.grd grdtrack [ERROR]: Must specify -G at least once gmtconvert [WARNING]: File <stdin> is empty! gmtconvert [WARNING]: No data records provided Sampling in geocoordinates with 50 meter pixels ... gmtinfo [ERROR]: No input data found! range: Subscript out of range. gmtmath [ERROR]: Operation "DIV" requires 2 operands gmtmath [ERROR]: Operation "MUL" requires 2 operands gmtinfo [ERROR]: No input data found! blockmedian [ERROR]: Unable to decode as a floating point number blockmedian [ERROR]: Option -I parsing failure. Correct syntax: -I<xinc>[+e|n][/<yinc>[+e|n]] Specify increment(s) and optionally append units or modifiers. For geographic regions in degrees you can optionally append units from this list: (d)egree [Default], (m)inute, (s)econd, m(e)ter, (f)oot, (k)ilometer, (M)ile, (n)autical mile, s(u)rvey foot. +e Adjust the region to fit increments [Adjust increment to fit domain]. +n Increment specifies the number of nodes instead. Then, the actual increments are calculated from the given domain and node- registration settings (see Appendix B for details). Note: If -R<grdfile> was used then -I (and -R and maybe -r) have been set; use -I to override those increments. blockmedian [ERROR]: Must specify -R option blockmedian [ERROR]: Option -I: Must specify positive increment(s) blockmedian [INFORMATION]: Provides 3, expects 3-column binary data xyz2grd [ERROR]: Unable to decode as a floating point number xyz2grd [ERROR]: Option -I parsing failure. Correct syntax: -I<xinc>[+e|n][/<yinc>[+e|n]] Specify increment(s) and optionally append units or modifiers. For geographic regions in degrees you can optionally append units from this list: (d)egree [Default], (m)inute, (s)econd, m(e)ter, (f)oot, (k)ilometer, (M)ile, (n)autical mile, s(u)rvey foot. +e Adjust the region to fit increments [Adjust increment to fit domain]. +n Increment specifies the number of nodes instead. Then, the actual increments are calculated from the given domain and node- registration settings (see Appendix B for details). Note: If -R<grdfile> was used then -I (and -R and maybe -r) have been set; use -I to override those increments. xyz2grd [ERROR]: Must specify -R option xyz2grd [ERROR]: Option -I: Must specify positive increment(s) grdinfo [ERROR]: Cannot find file corr.grd grdinfo [ERROR]: Must specify one or more input files makecpt [ERROR]: Option T: Must specify valid min/max/inc[<unit>|+n] option grdinfo [ERROR]: Cannot find file phasefilt_ll.grd grdinfo [ERROR]: Must specify one or more input files gmtmath [ERROR]: Operation "INV" requires 1 operands grdimage [ERROR]: Cannot find file phasefilt_ll.grd Make phasefilt_ll.kml and phasefilt_ll.png psconvert [ERROR]: Option -E: No argument provided rm: No match. grdinfo [ERROR]: Cannot find file phasefilt_mask_ll.grd grdinfo [ERROR]: Must specify one or more input files gmtmath [ERROR]: Operation "INV" requires 1 operands grdimage [ERROR]: Cannot find file phasefilt_mask_ll.grd Make phasefilt_mask_ll.kml and phasefilt_mask_ll.png psconvert [ERROR]: Option -E: No argument provided rm: No match. grdinfo [ERROR]: Cannot find file corr_ll.grd grdinfo [ERROR]: Must specify one or more input files gmtmath [ERROR]: Operation "INV" requires 1 operands grdimage [ERROR]: Cannot find file corr_ll.grd Make corr_ll.kml and corr_ll.png psconvert [ERROR]: Option -E: No argument provided rm: No match. GEOCODE END rm: No match. DInSAR processing completed successfully.
05-14
PowerShell 7 环境已加载 (版本: 7.5.2) PowerShell 7 环境已加载 (版本: 7.5.2) PS C:\Users\Administrator\Desktop> cd E:\PyTorch_Build\pytorch PS E:\PyTorch_Build\pytorch> python -m venv rtx5070_env PS E:\PyTorch_Build\pytorch> .\rtx5070_env\Scripts\activate (rtx5070_env) PS E:\PyTorch_Build\pytorch> # 1. 运行修复脚本 (rtx5070_env) PS E:\PyTorch_Build\pytorch> .\complete_install_fix.ps1 尝试从备用源下载: https://pilotfiber.dl.sourceforge.net/project/openblas/v0.3.27/OpenBLAS-0.3.27-x64.zip ❌ 下载失败: The variable '$localZip' cannot be retrieved because it has not been set. 尝试从备用源下载: https://cfhcable.dl.sourceforge.net/project/openblas/v0.3.27/OpenBLAS-0.3.27-x64.zip ❌ 下载失败: The variable '$localZip' cannot be retrieved because it has not been set. 尝试从备用源下载: https://superb-sea2.dl.sourceforge.net/project/openblas/v0.3.27/OpenBLAS-0.3.27-x64.zip ❌ 下载失败: The variable '$localZip' cannot be retrieved because it has not been set. 尝试从GitHub直接下载... InvalidOperation: E:\PyTorch_Build\pytorch\complete_install_fix.ps1:25 Line | 25 | Invoke-WebRequest -Uri $githubRaw -OutFile $localZip | ~~~~~~~~~ | The variable '$localZip' cannot be retrieved because it has not been set. InvalidOperation: E:\PyTorch_Build\pytorch\complete_install_fix.ps1:35 Line | 35 | [System.IO.Compression.ZipFile]::ExtractToDirectory($localZip, $newIn … | ~~~~~~~~~ | The variable '$localZip' cannot be retrieved because it has not been set. Find-BLASBinary: E:\PyTorch_Build\pytorch\complete_install_fix.ps1:38 Line | 38 | $blasFile = Find-BLASBinary -SearchPath $newInstallPath | ~~~~~~~~~~~~~~~ | The term 'Find-BLASBinary' is not recognized as a name of a cmdlet, function, script file, or executable | program. Check the spelling of the name, or if a path was included, verify that the path is correct and try | again. ⚠️ 二进制分发版无效,尝试从源码编译... Chocolatey v2.5.0 3 validations performed. 2 success(es), 1 warning(s), and 0 error(s). Validation Warnings: - A pending system reboot request has been detected, however, this is being ignored due to the current Chocolatey configuration. If you want to halt when this occurs, then either set the global feature using: choco feature enable --name="exitOnRebootDetected" or pass the option --exit-when-reboot-detected. Installing the following packages: cmake By installing, you accept licenses for the packages. Downloading package from source 'https://community.chocolatey.org/api/v2/' Progress: Downloading cmake.install 4.1.1... 100% cmake.install v4.1.1 [Approved] cmake.install package files install completed. Performing other installation steps. Installing 64-bit cmake.install... cmake.install has been installed. The install of cmake.install was successful. Software installed as 'msi', install location is likely default. Downloading package from source 'https://community.chocolatey.org/api/v2/' Progress: Downloading cmake 4.1.1... 100% cmake v4.1.1 [Approved] cmake package files install completed. Performing other installation steps. The install of cmake was successful. Deployed to 'C:\ProgramData\chocolatey\lib\cmake' Chocolatey installed 2/2 packages. See the log for details (C:\ProgramData\chocolatey\logs\chocolatey.log). Chocolatey v2.5.0 3 validations performed. 2 success(es), 1 warning(s), and 0 error(s). Validation Warnings: - A pending system reboot request has been detected, however, this is being ignored due to the current Chocolatey configuration. If you want to halt when this occurs, then either set the global feature using: choco feature enable --name="exitOnRebootDetected" or pass the option --exit-when-reboot-detected. Installing the following packages: mingw By installing, you accept licenses for the packages. Downloading package from source 'https://community.chocolatey.org/api/v2/' Progress: Downloading mingw 13.2.0... 100% mingw v13.2.0 [Approved] mingw package files install completed. Performing other installation steps. Downloading mingw 64 bit from 'https://github.com/niXman/mingw-builds-binaries/releases/download/13.2.0-rt_v11-rev0/x86_64-13.2.0-release-posix-seh-ucrt-rt_v11-rev0.7z' Progress: 100% - Completed download of C:\Users\Administrator\AppData\Local\Temp\chocolatey\mingw\13.2.0\x86_64-13.2.0-release-posix-seh-ucrt-rt_v11-rev0.7z (69.57 MB). Download of x86_64-13.2.0-release-posix-seh-ucrt-rt_v11-rev0.7z (69.57 MB) completed. Hashes match. Extracting C:\Users\Administrator\AppData\Local\Temp\chocolatey\mingw\13.2.0\x86_64-13.2.0-release-posix-seh-ucrt-rt_v11-rev0.7z to C:\ProgramData\chocolatey\lib\mingw\tools\install... C:\ProgramData\chocolatey\lib\mingw\tools\install Testing path: C:\ProgramData\mingw64\mingw32\bin Testing path: C:\ProgramData\mingw64\mingw64\bin PATH environment variable does not have C:\ProgramData\mingw64\mingw64\bin in it. Adding... Environment Vars (like PATH) have changed. Close/reopen your shell to see the changes (or in powershell/cmd.exe just type `refreshenv`). The install of mingw was successful. Deployed to 'C:\ProgramData\chocolatey\lib\mingw\tools\install' Chocolatey installed 1/1 packages. See the log for details (C:\ProgramData\chocolatey\logs\chocolatey.log). Cloning into 'OpenBLAS'... remote: Enumerating objects: 81610, done. remote: Counting objects: 100% (42/42), done. remote: Compressing objects: 100% (26/26), done. remote: Total 81610 (delta 15), reused 21 (delta 6), pack-reused 81568 (from 2) Receiving objects: 100% (81610/81610), 59.50 MiB | 6.03 MiB/s, done. Resolving deltas: 100% (65602/65602), done. Updating files: 100% (12531/12531), done. Directory: E:\PyTorch_Build\pytorch\OpenBLAS Mode LastWriteTime Length Name ---- ------------- ------ ---- d---- 2025/9/2 0:29 build CMake Error: CMake was unable to find a build program corresponding to "MinGW Makefiles". CMAKE_MAKE_PROGRAM is not set. You probably need to select a different build tool. CMake Error: CMAKE_C_COMPILER not set, after EnableLanguage CMake Error: CMAKE_ASM_COMPILER not set, after EnableLanguage -- Configuring incomplete, errors occurred! mingw32-make: E:\PyTorch_Build\pytorch\build_openblas_from_source.ps1:14 Line | 14 | mingw32-make -j8 | ~~~~~~~~~~~~ | The term 'mingw32-make' is not recognized as a name of a cmdlet, function, script file, or executable program. | Check the spelling of the name, or if a path was included, verify that the path is correct and try again. mingw32-make: E:\PyTorch_Build\pytorch\build_openblas_from_source.ps1:15 Line | 15 | mingw32-make install | ~~~~~~~~~~~~ | The term 'mingw32-make' is not recognized as a name of a cmdlet, function, script file, or executable program. | Check the spelling of the name, or if a path was included, verify that the path is correct and try again. (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> # 2. 验证环境 (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> Write-Host "当前OpenBLAS_HOME: $env:OpenBLAS_HOME" 当前OpenBLAS_HOME: E:\Libs\OpenBLAS_Build (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> # 3. 检查文件 (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> dir "$env:OpenBLAS_HOME" -Recurse | Where-Object { $_.Name -like "*blas*" } (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> # 4. 测试PyTorch配置 (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> python -c "import os; print('USE_OPENBLAS:', os.environ.get('USE_OPENBLAS')); print('路径验证:', os.path.exists(os.path.join(os.environ['OpenBLAS_HOME'], 'bin', 'openblas.dll')))" USE_OPENBLAS: None 路径验证: False (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> # 1. 保存修复后的脚本 (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> # 将上述两个脚本保存为: (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> # - fixed_complete_install_fix.ps1 (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> # - fixed_build_openblas_from_source.ps1 (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> # 2. 启动自动安装流程 (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> .\fixed_complete_install_fix.ps1 .\fixed_complete_install_fix.ps1: The term '.\fixed_complete_install_fix.ps1' is not recognized as a name of a cmdlet, function, script file, or executable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> # 3. 设置电源选项(防止休眠) (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> powercfg -change -standby-timeout-ac 0 (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> powercfg -change -hibernate-timeout-ac 0 (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> # 4. 启用远程监控(可选) (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> $webhook = "https://webhook.site/your-unique-url" (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> $status = "开始编译OpenBLAS" (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build> Invoke-RestMethod -Uri $webhook -Method Post -Body $status Invoke-RestMethod: Error: Token "your-unique-url" not found - Webhook.site Webhook.site Error Token "your-unique-url" not found The URL was deleted, or expired automatically. To avoid URLs expiring automatically, you can upgrade to a Webhook.site subscription. Upgrade Now Home   Control Panel   Create Account   Contact Support   Login (rtx5070_env) PS E:\PyTorch_Build\pytorch\OpenBLAS\build>
09-03
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值