experience

cmd下Oracle命令:
lsnrctl start
net start oracleserviceorcl
tnsping ORCL.CHINA

---------------------------------------------------
文本框中只输入数字:
onkeyup="this.value=this.value.replace(/\D/g,'')"
---------------------------------------------------
mysql导文件:
mysqlimport -pAdmin@123 dbname --fields-terminated-by=',' '/tmp/T_INTERNETUSERINFOSYNC_TMP.txt'
mysql -uroot -pAdmin@123 dbname < /root/hour.txt
--------------------------------------------------------
IP点阵式转整数:
SELECT * FROM TABLE WHERE ip= INET_ATON('222.244.230.17')
IP整数转点阵式:
SELECT * FROM TABLE WHERE ip= INET_NTOA(136524587)
--------------------------------------------------------
mysql load数据
LOAD DATA LOCAL INFILE 'E://cus.csv' INTO TABLE T_BIG_CUS FIELDS TERMINATED BY ',' (字段1,字段2)
---------------------------------------------------------
SHOW FULL PROCESSLIST
SHOW PROCESSLIST

远程拷贝:
scp root@128.20.59.6:/root/hour.txt /root/

jdk查看线程命令
jstack -l 4256

linux环境下添加路由:
查询路由信息:route
route add –net 128.18.0.0 255.255.0.0 MASK gw 128.20.0.0

window route add 128.18.38.0 mask 255.255.255.0 128.20.1.1
unix route add -net 128.18.38.0 netmask 255.255.255.0 gw 128.20.1.1

killall -9 httpd
统计进程数:
ps -ef | grep bin/httpd | wc -l

查看达到的并发数:
webbench -c 1000 -t 30 http://ip:8080/index.html

linux增加可执行权限:chmod +x jpda.sh

添加路由:
route print
route add 128.18.0.0 mask 255.255.0.0 128.20.1.1 -p

拷贝数据:
mysqldump -uroot -pAdmin@123 "--where=AREA_ID = 0 AND TYPE = 255" db_name T_TRAFFIC | mysql -uroot -pAdmin@123 -h128.20.20.103 db_name

unix时间:
SELECT UNIX_TIMESTAMP('2011-05-30 22:00:00')
SELECT FROM_UNIXTIME(1306892522)
### Experience Replay 概念 Experience Replay 是一种用于提高强化学习算法稳定性和效率的技术。通过存储代理在过去经历过的状态、动作、奖励和下一个状态的元组到一个称为回放缓冲区的数据结构中,在训练过程中可以随机抽取这些经验样本进行更新,而不是仅依赖最近的经验数据[^2]。 这种方法有助于打破数据之间的关联性,使得模型能够更有效地泛化所学的知识。此外,它还允许重复利用历史数据,从而提高了数据使用的效率并减少了对新交互的需求。 ### 实现方法 为了实现 Experience Replay 功能,通常会创建一个循环缓冲区来保存一定数量的历史体验样本。当这个缓冲区满了之后,新的样本将会替换掉最旧的那个。下面是一个简单的 Python 版本的 Experience Replay 缓存类: ```python import random from collections import deque, namedtuple class ReplayBuffer: """Fixed-size buffer to store experience tuples.""" def __init__(self, buffer_size, batch_size): self.memory = deque(maxlen=buffer_size) self.batch_size = batch_size self.experience = namedtuple("Experience", field_names=["state", "action", "reward", "next_state", "done"]) def add(self, state, action, reward, next_state, done): e = self.experience(state, action, reward, next_state, done) self.memory.append(e) def sample(self): experiences = random.sample(self.memory, k=self.batch_size) states = torch.from_numpy(np.vstack([e.state for e in experiences if e is not None])).float().to(device) actions = torch.from_numpy(np.vstack([e.action for e in experiences if e is not None])).long().to(device) rewards = torch.from_numpy(np.vstack([e.reward for e in experiences if e is not None])).float().to(device) next_states = torch.from_numpy(np.vstack([e.next_state for e in experiences if e is not None])).float().to(device) dones = torch.from_numpy(np.vstack([e.done for e in experiences if e is not None]).astype(np.uint8)).float().to(device) return (states, actions, rewards, next_states, dones) def __len__(self): return len(self.memory) ``` 这段代码定义了一个 `ReplayBuffer` 类,该类维护着一个固定大小的记忆列表,并提供了向其中添加新条目以及从中采样一批条目的功能。注意这里假设使用 PyTorch 进行张量操作;如果采用其他框架,则需相应调整转换逻辑。
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值