新增网络socket的系统调用接口 (#247)

1.修复spinlock忘记恢复rflags的问题
2.WaitQueue增加wakeup_all的功能
3.完善tcp,udp,raw socket
4.把PollStatus结构体改为使用bitflags
5.新增iovec结构体
6.完成网络的系统调用
7.在bootstrap里面添加dnsmasq bridge-utils iptables

---------

Co-authored-by: guanjinquan <1666320330@qq.com>
This commit is contained in:
login 2023-04-19 18:05:02 +08:00 committed by GitHub
parent 8fd71f2772
commit cde5492f72
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
33 changed files with 2535 additions and 240 deletions

1
.gitignore vendored
View File

@ -3,6 +3,7 @@
!.gitkeep !.gitkeep
DragonOS.iso DragonOS.iso
.idea/ .idea/
/tmp/
kernel/kernel kernel/kernel
.DS_Store .DS_Store

View File

@ -108,9 +108,7 @@ impl IndexNode for LockedAhciInode {
} }
fn poll(&self) -> Result<PollStatus, SystemError> { fn poll(&self) -> Result<PollStatus, SystemError> {
return Ok(PollStatus { return Ok(PollStatus::READ | PollStatus::WRITE);
flags: PollStatus::READ_MASK | PollStatus::WRITE_MASK,
});
} }
/// 读设备 - 应该调用设备的函数读写,而不是通过文件系统读写 /// 读设备 - 应该调用设备的函数读写,而不是通过文件系统读写

View File

@ -150,9 +150,7 @@ impl IndexNode for LockedPS2KeyBoardInode {
} }
fn poll(&self) -> Result<PollStatus, SystemError> { fn poll(&self) -> Result<PollStatus, SystemError> {
return Ok(PollStatus { return Ok(PollStatus::READ);
flags: PollStatus::READ_MASK,
});
} }
fn metadata(&self) -> Result<Metadata, SystemError> { fn metadata(&self) -> Result<Metadata, SystemError> {

View File

@ -10,7 +10,7 @@ use virtio_drivers::{device::net::VirtIONet, transport::Transport};
use crate::{ use crate::{
driver::{virtio::virtio_impl::HalImpl, Driver}, driver::{virtio::virtio_impl::HalImpl, Driver},
kdebug, kerror, kinfo, kerror, kinfo,
libs::spinlock::SpinLock, libs::spinlock::SpinLock,
net::{generate_iface_id, NET_DRIVERS}, net::{generate_iface_id, NET_DRIVERS},
syscall::SystemError, syscall::SystemError,

View File

@ -245,8 +245,9 @@ void do_page_fault(struct pt_regs *regs, unsigned long error_code)
printk_color(RED, BLACK, "CR2:%#018lx\n", cr2); printk_color(RED, BLACK, "CR2:%#018lx\n", cr2);
traceback(regs); traceback(regs);
current_pcb->state = PROC_STOPPED; process_do_exit(-1);
sched(); // current_pcb->state = PROC_STOPPED;
// sched();
} }
// 15 Intel保留请勿使用 // 15 Intel保留请勿使用

View File

@ -465,9 +465,7 @@ impl IndexNode for LockedDevFSInode {
return Err(SystemError::EISDIR); return Err(SystemError::EISDIR);
} }
return Ok(PollStatus { return Ok(PollStatus::READ | PollStatus::WRITE);
flags: PollStatus::READ_MASK | PollStatus::WRITE_MASK,
});
} }
/// 读设备 - 应该调用设备的函数读写,而不是通过文件系统读写 /// 读设备 - 应该调用设备的函数读写,而不是通过文件系统读写

View File

@ -102,9 +102,7 @@ impl IndexNode for LockedNullInode {
} }
fn poll(&self) -> Result<PollStatus, SystemError> { fn poll(&self) -> Result<PollStatus, SystemError> {
return Ok(PollStatus { return Ok(PollStatus::READ | PollStatus::WRITE);
flags: PollStatus::READ_MASK | PollStatus::WRITE_MASK,
});
} }
/// 读设备 - 应该调用设备的函数读写,而不是通过文件系统读写 /// 读设备 - 应该调用设备的函数读写,而不是通过文件系统读写

View File

@ -102,9 +102,7 @@ impl IndexNode for LockedZeroInode {
} }
fn poll(&self) -> Result<PollStatus, SystemError> { fn poll(&self) -> Result<PollStatus, SystemError> {
return Ok(PollStatus { return Ok(PollStatus::READ | PollStatus::WRITE);
flags: PollStatus::READ_MASK | PollStatus::WRITE_MASK,
});
} }
/// 读设备 - 应该调用设备的函数读写,而不是通过文件系统读写 /// 读设备 - 应该调用设备的函数读写,而不是通过文件系统读写

View File

@ -1411,9 +1411,7 @@ impl IndexNode for LockedFATInode {
return Err(SystemError::EISDIR); return Err(SystemError::EISDIR);
} }
return Ok(PollStatus { return Ok(PollStatus::READ | PollStatus::WRITE);
flags: PollStatus::READ_MASK | PollStatus::WRITE_MASK,
});
} }
fn create( fn create(

View File

@ -425,9 +425,7 @@ impl IndexNode for LockedProcFSInode {
return Err(SystemError::EISDIR); return Err(SystemError::EISDIR);
} }
return Ok(PollStatus { return Ok(PollStatus::READ);
flags: PollStatus::READ_MASK,
});
} }
fn fs(&self) -> Arc<dyn FileSystem> { fn fs(&self) -> Arc<dyn FileSystem> {

View File

@ -186,9 +186,7 @@ impl IndexNode for LockedRamFSInode {
return Err(SystemError::EISDIR); return Err(SystemError::EISDIR);
} }
return Ok(PollStatus { return Ok(PollStatus::READ | PollStatus::WRITE);
flags: PollStatus::READ_MASK | PollStatus::WRITE_MASK,
});
} }
fn fs(&self) -> Arc<dyn FileSystem> { fn fs(&self) -> Arc<dyn FileSystem> {

View File

@ -290,6 +290,12 @@ impl File {
return Some(res); return Some(res);
} }
/// @brief 获取文件的类型
#[inline]
pub fn file_type(&self) -> FileType {
return self.file_type;
}
} }
impl Drop for File { impl Drop for File {

View File

@ -3,14 +3,14 @@
pub mod core; pub mod core;
pub mod file; pub mod file;
pub mod mount; pub mod mount;
mod syscall; pub mod syscall;
mod utils; mod utils;
use ::core::{any::Any, fmt::Debug}; use ::core::{any::Any, fmt::Debug};
use alloc::{string::String, sync::Arc, vec::Vec}; use alloc::{string::String, sync::Arc, vec::Vec};
use crate::{syscall::SystemError, time::TimeSpec}; use crate::{libs::casting::DowncastArc, syscall::SystemError, time::TimeSpec};
use self::{core::generate_inode_id, file::FileMode}; use self::{core::generate_inode_id, file::FileMode};
pub use self::{core::ROOT_INODE, file::FilePrivateData, mount::MountFS}; pub use self::{core::ROOT_INODE, file::FilePrivateData, mount::MountFS};
@ -36,6 +36,8 @@ pub enum FileType {
Pipe, Pipe,
/// 符号链接 /// 符号链接
SymLink, SymLink,
/// 套接字
Socket,
} }
/* these are defined by POSIX and also present in glibc's dirent.h */ /* these are defined by POSIX and also present in glibc's dirent.h */
@ -68,20 +70,18 @@ impl FileType {
FileType::CharDevice => DT_CHR, FileType::CharDevice => DT_CHR,
FileType::Pipe => DT_FIFO, FileType::Pipe => DT_FIFO,
FileType::SymLink => DT_LNK, FileType::SymLink => DT_LNK,
FileType::Socket => DT_SOCK,
}; };
} }
} }
/// @brief inode的状态由poll方法返回 bitflags! {
#[derive(Debug, Default, PartialEq)] /// @brief inode的状态由poll方法返回
pub struct PollStatus { pub struct PollStatus: u8 {
pub flags: u8, const WRITE = 1u8 << 0;
} const READ = 1u8 << 1;
const ERROR = 1u8 << 2;
impl PollStatus { }
pub const WRITE_MASK: u8 = (1u8 << 0);
pub const READ_MASK: u8 = (1u8 << 1);
pub const ERR_MASK: u8 = (1u8 << 2);
} }
pub trait IndexNode: Any + Sync + Send + Debug { pub trait IndexNode: Any + Sync + Send + Debug {
@ -336,6 +336,12 @@ pub trait IndexNode: Any + Sync + Send + Debug {
} }
} }
impl DowncastArc for dyn IndexNode {
fn as_any_arc(self: Arc<Self>) -> Arc<dyn Any> {
self
}
}
impl dyn IndexNode { impl dyn IndexNode {
/// @brief 将当前Inode转换为一个具体的结构体类型由T指定 /// @brief 将当前Inode转换为一个具体的结构体类型由T指定
/// 如果类型正确则返回Some,否则返回None /// 如果类型正确则返回Some,否则返回None
@ -482,6 +488,27 @@ pub struct Metadata {
pub raw_dev: usize, pub raw_dev: usize,
} }
impl Default for Metadata {
fn default() -> Self {
return Self {
dev_id: 0,
inode_id: 0,
size: 0,
blk_size: 0,
blocks: 0,
atime: TimeSpec::default(),
mtime: TimeSpec::default(),
ctime: TimeSpec::default(),
file_type: FileType::File,
mode: 0,
nlinks: 1,
uid: 0,
gid: 0,
raw_dev: 0,
};
}
}
/// @brief 所有文件系统都应该实现的trait /// @brief 所有文件系统都应该实现的trait
pub trait FileSystem: Any + Sync + Send + Debug { pub trait FileSystem: Any + Sync + Send + Debug {
/// @brief 获取当前文件系统的root inode的指针 /// @brief 获取当前文件系统的root inode的指针

View File

@ -1,6 +1,6 @@
use core::ffi::{c_char, CStr}; use core::ffi::{c_char, CStr};
use alloc::{boxed::Box, string::ToString}; use alloc::{boxed::Box, string::ToString, vec::Vec};
use crate::{ use crate::{
arch::asm::{current::current_pcb, ptrace::user_mode}, arch::asm::{current::current_pcb, ptrace::user_mode},
@ -435,3 +435,106 @@ pub extern "C" fn sys_dup2(regs: &pt_regs) -> u64 {
return r.unwrap_err().to_posix_errno() as u64; return r.unwrap_err().to_posix_errno() as u64;
} }
} }
#[repr(C)]
#[derive(Debug, Clone, Copy)]
pub struct IoVec {
/// 缓冲区的起始地址
pub iov_base: *mut u8,
/// 缓冲区的长度
pub iov_len: usize,
}
/// 用于存储多个来自用户空间的IoVec
///
/// 由于目前内核中的文件系统还不支持分散读写所以暂时只支持将用户空间的IoVec聚合成一个缓冲区然后进行操作。
/// TODO支持分散读写
#[derive(Debug)]
pub struct IoVecs(Vec<&'static mut [u8]>);
impl IoVecs {
/// 从用户空间的IoVec中构造IoVecs
///
/// @param iov 用户空间的IoVec
/// @param iovcnt 用户空间的IoVec的数量
/// @param readv 是否为readv系统调用
///
/// @return 构造成功返回IoVecs否则返回错误码
pub unsafe fn from_user(
iov: *const IoVec,
iovcnt: usize,
_readv: bool,
) -> Result<Self, SystemError> {
// 检查iov指针所在空间是否合法
if !verify_area(
iov as usize as u64,
(iovcnt * core::mem::size_of::<IoVec>()) as u64,
) {
return Err(SystemError::EFAULT);
}
// 将用户空间的IoVec转换为引用注意这里的引用是静态的因为用户空间的IoVec不会被释放
let iovs: &[IoVec] = core::slice::from_raw_parts(iov, iovcnt);
let mut slices: Vec<&mut [u8]> = vec![];
slices.reserve(iovs.len());
for iov in iovs.iter() {
if iov.iov_len == 0 {
continue;
}
if !verify_area(iov.iov_base as usize as u64, iov.iov_len as u64) {
return Err(SystemError::EFAULT);
}
slices.push(core::slice::from_raw_parts_mut(iov.iov_base, iov.iov_len));
}
return Ok(Self(slices));
}
/// @brief 将IoVecs中的数据聚合到一个缓冲区中
///
/// @return 返回聚合后的缓冲区
pub fn gather(&self) -> Vec<u8> {
let mut buf = Vec::new();
for slice in self.0.iter() {
buf.extend_from_slice(slice);
}
return buf;
}
/// @brief 将给定的数据分散写入到IoVecs中
pub fn scatter(&mut self, data: &[u8]) {
let mut data: &[u8] = data;
for slice in self.0.iter_mut() {
let len = core::cmp::min(slice.len(), data.len());
if len == 0 {
continue;
}
slice[..len].copy_from_slice(&data[..len]);
data = &data[len..];
}
}
/// @brief 创建与IoVecs等长的缓冲区
///
/// @param set_len 是否设置返回的Vec的len。
/// 如果为true则返回的Vec的len为所有IoVec的长度之和;
/// 否则返回的Vec的len为0capacity为所有IoVec的长度之和.
///
/// @return 返回创建的缓冲区
pub fn new_buf(&self, set_len: bool) -> Vec<u8> {
let total_len: usize = self.0.iter().map(|slice| slice.len()).sum();
let mut buf: Vec<u8> = Vec::with_capacity(total_len);
if set_len {
unsafe {
buf.set_len(total_len);
}
}
return buf;
}
}

View File

@ -6,6 +6,7 @@
#![feature(panic_info_message)] #![feature(panic_info_message)]
#![feature(drain_filter)] // 允许Vec的drain_filter特性 #![feature(drain_filter)] // 允许Vec的drain_filter特性
#![feature(c_void_variant)] #![feature(c_void_variant)]
#![feature(trait_upcasting)]
#[allow(non_upper_case_globals)] #[allow(non_upper_case_globals)]
#[allow(non_camel_case_types)] #[allow(non_camel_case_types)]
#[allow(non_snake_case)] #[allow(non_snake_case)]
@ -99,6 +100,9 @@ pub fn panic(info: &PanicInfo) -> ! {
#[no_mangle] #[no_mangle]
pub extern "C" fn __rust_demo_func() -> i32 { pub extern "C" fn __rust_demo_func() -> i32 {
printk_color!(GREEN, BLACK, "__rust_demo_func()\n"); printk_color!(GREEN, BLACK, "__rust_demo_func()\n");
net_init().expect("Failed to init network"); let r = net_init();
if r.is_err() {
kwarn!("net_init() failed: {:?}", r.err().unwrap());
}
return 0; return 0;
} }

View File

@ -53,7 +53,7 @@ use alloc::sync::Arc;
/// assert!(a_arc2.is_some()); /// assert!(a_arc2.is_some());
/// } /// }
/// ``` /// ```
trait DowncastArc: Any + Send + Sync { pub trait DowncastArc: Any + Send + Sync {
/// 请在具体类型中实现这个函数返回self /// 请在具体类型中实现这个函数返回self
fn as_any_arc(self: Arc<Self>) -> Arc<dyn Any>; fn as_any_arc(self: Arc<Self>) -> Arc<dyn Any>;

View File

@ -244,6 +244,10 @@ impl<T> DerefMut for SpinLockGuard<'_, T> {
/// @brief 为SpinLockGuard实现Drop方法那么一旦守卫的生命周期结束就会自动释放自旋锁避免了忘记放锁的情况 /// @brief 为SpinLockGuard实现Drop方法那么一旦守卫的生命周期结束就会自动释放自旋锁避免了忘记放锁的情况
impl<T> Drop for SpinLockGuard<'_, T> { impl<T> Drop for SpinLockGuard<'_, T> {
fn drop(&mut self) { fn drop(&mut self) {
self.lock.lock.unlock(); if self.flag != 0 {
self.lock.lock.unlock_irqrestore(&self.flag);
} else {
self.lock.lock.unlock();
}
} }
} }

View File

@ -1,5 +1,5 @@
#![allow(dead_code)] #![allow(dead_code)]
use alloc::collections::LinkedList; use alloc::{collections::LinkedList, vec::Vec};
use crate::{ use crate::{
arch::{asm::current::current_pcb, sched::sched}, arch::{asm::current::current_pcb, sched::sched},
@ -127,6 +127,33 @@ impl WaitQueue {
} }
} }
/// @brief 唤醒在队列中,符合条件的所有进程。
///
/// @param state 用于判断的state如果队列中第一个进程的state与它进行and操作之后结果不为0,则唤醒这个进程。
pub fn wakeup_all(&self, state: u64) {
let mut guard: SpinLockGuard<InnerWaitQueue> = self.0.lock_irqsave();
// 如果队列为空,则返回
if guard.wait_list.is_empty() {
return;
}
let mut to_push_back: Vec<&mut process_control_block> = Vec::new();
// 如果队列头部的pcb的state与给定的state相与结果不为0则唤醒
while let Some(to_wakeup) = guard.wait_list.pop_front() {
if (to_wakeup.state & state) != 0 {
unsafe {
process_wakeup(to_wakeup);
}
} else {
to_push_back.push(to_wakeup);
}
}
for to_wakeup in to_push_back {
guard.wait_list.push_back(to_wakeup);
}
}
/// @brief 获得当前等待队列的大小 /// @brief 获得当前等待队列的大小
pub fn len(&self) -> usize { pub fn len(&self) -> usize {
return self.0.lock().wait_list.len(); return self.0.lock().wait_list.len();

View File

@ -50,6 +50,6 @@ unsafe impl GlobalAlloc for KernelAllocator {
/// 内存分配错误处理函数 /// 内存分配错误处理函数
#[alloc_error_handler] #[alloc_error_handler]
pub fn global_alloc_err_handler(_layout: Layout) -> ! { pub fn global_alloc_err_handler(layout: Layout) -> ! {
panic!("global_alloc_error"); panic!("global_alloc_error, layout: {:?}", layout);
} }

View File

@ -5,7 +5,7 @@ use core::{
use alloc::{boxed::Box, collections::BTreeMap, sync::Arc}; use alloc::{boxed::Box, collections::BTreeMap, sync::Arc};
use crate::{driver::net::NetDriver, libs::rwlock::RwLock, syscall::SystemError}; use crate::{driver::net::NetDriver, kwarn, libs::rwlock::RwLock, syscall::SystemError};
use smoltcp::wire::IpEndpoint; use smoltcp::wire::IpEndpoint;
use self::socket::SocketMetadata; use self::socket::SocketMetadata;
@ -13,6 +13,7 @@ use self::socket::SocketMetadata;
pub mod endpoints; pub mod endpoints;
pub mod net_core; pub mod net_core;
pub mod socket; pub mod socket;
pub mod syscall;
lazy_static! { lazy_static! {
/// @brief 所有网络接口的列表 /// @brief 所有网络接口的列表
@ -28,11 +29,28 @@ pub fn generate_iface_id() -> usize {
} }
/// @brief 用于指定socket的关闭类型 /// @brief 用于指定socket的关闭类型
#[derive(Debug, Clone)] /// 参考https://pubs.opengroup.org/onlinepubs/9699919799/functions/shutdown.html
#[derive(Debug, Clone, Copy, PartialEq, Eq, FromPrimitive, ToPrimitive)]
pub enum ShutdownType { pub enum ShutdownType {
ShutRd, // Disables further receive operations. ShutRd = 0, // Disables further receive operations.
ShutWr, // Disables further send operations. ShutWr = 1, // Disables further send operations.
ShutRdwr, // Disables further send and receive operations. ShutRdwr = 2, // Disables further send and receive operations.
}
impl TryFrom<i32> for ShutdownType {
type Error = SystemError;
fn try_from(value: i32) -> Result<Self, Self::Error> {
use num_traits::FromPrimitive;
<Self as FromPrimitive>::from_i32(value).ok_or(SystemError::EINVAL)
}
}
impl Into<i32> for ShutdownType {
fn into(self) -> i32 {
use num_traits::ToPrimitive;
<Self as ToPrimitive>::to_i32(&self).unwrap()
}
} }
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
@ -40,7 +58,7 @@ pub enum Endpoint {
/// 链路层端点 /// 链路层端点
LinkLayer(endpoints::LinkLayerEndpoint), LinkLayer(endpoints::LinkLayerEndpoint),
/// 网络层端点 /// 网络层端点
Ip(IpEndpoint), Ip(Option<IpEndpoint>),
// todo: 增加NetLink机制后增加NetLink端点 // todo: 增加NetLink机制后增加NetLink端点
} }
@ -51,7 +69,7 @@ pub trait Socket: Sync + Send + Debug {
/// ///
/// @return - 成功:(返回读取的数据的长度,读取数据的端点). /// @return - 成功:(返回读取的数据的长度,读取数据的端点).
/// - 失败:错误码 /// - 失败:错误码
fn read(&self, buf: &mut [u8]) -> Result<(usize, Endpoint), SystemError>; fn read(&self, buf: &mut [u8]) -> (Result<usize, SystemError>, Endpoint);
/// @brief 向socket中写入数据。如果socket是阻塞的那么直到写入的数据全部写入socket中才返回 /// @brief 向socket中写入数据。如果socket是阻塞的那么直到写入的数据全部写入socket中才返回
/// ///
@ -80,7 +98,7 @@ pub trait Socket: Sync + Send + Debug {
/// @param endpoint 要绑定的端点 /// @param endpoint 要绑定的端点
/// ///
/// @return 返回绑定是否成功 /// @return 返回绑定是否成功
fn bind(&self, _endpoint: Endpoint) -> Result<(), SystemError> { fn bind(&mut self, _endpoint: Endpoint) -> Result<(), SystemError> {
return Err(SystemError::ENOSYS); return Err(SystemError::ENOSYS);
} }
@ -106,7 +124,7 @@ pub trait Socket: Sync + Send + Debug {
/// @brief 对应于POSIX的accept函数用于接受连接 /// @brief 对应于POSIX的accept函数用于接受连接
/// ///
/// @param endpoint 用于返回连接的端点 /// @param endpoint 对端的端点
/// ///
/// @return 返回接受连接是否成功 /// @return 返回接受连接是否成功
fn accept(&mut self) -> Result<(Box<dyn Socket>, Endpoint), SystemError> { fn accept(&mut self) -> Result<(Box<dyn Socket>, Endpoint), SystemError> {
@ -164,6 +182,23 @@ pub trait Socket: Sync + Send + Debug {
fn metadata(&self) -> Result<SocketMetadata, SystemError>; fn metadata(&self) -> Result<SocketMetadata, SystemError>;
fn box_clone(&self) -> Box<dyn Socket>; fn box_clone(&self) -> Box<dyn Socket>;
/// @brief 设置socket的选项
///
/// @param level 选项的层次
/// @param optname 选项的名称
/// @param optval 选项的值
///
/// @return 返回设置是否成功, 如果不支持该选项返回ENOSYS
fn setsockopt(
&self,
_level: usize,
_optname: usize,
_optval: &[u8],
) -> Result<(), SystemError> {
kwarn!("setsockopt is not implemented");
return Err(SystemError::ENOSYS);
}
} }
impl Clone for Box<dyn Socket> { impl Clone for Box<dyn Socket> {

View File

@ -1,15 +1,42 @@
use alloc::{boxed::Box, collections::BTreeMap, sync::Arc};
use smoltcp::{socket::dhcpv4, wire}; use smoltcp::{socket::dhcpv4, wire};
use crate::{kdebug, kinfo, net::NET_DRIVERS, syscall::SystemError}; use crate::{
driver::net::NetDriver,
kdebug, kinfo, kwarn,
libs::rwlock::RwLockReadGuard,
net::NET_DRIVERS,
syscall::SystemError,
time::timer::{next_n_ms_timer_jiffies, Timer, TimerFunction},
};
use super::socket::{SOCKET_SET, SOCKET_WAITQUEUE};
/// The network poll function, which will be called by timer.
///
/// The main purpose of this function is to poll all network interfaces.
struct NetWorkPollFunc();
impl TimerFunction for NetWorkPollFunc {
fn run(&mut self) {
poll_ifaces_try_lock(10).ok();
let next_time = next_n_ms_timer_jiffies(10);
let timer = Timer::new(Box::new(NetWorkPollFunc()), next_time);
timer.activate();
}
}
pub fn net_init() -> Result<(), SystemError> { pub fn net_init() -> Result<(), SystemError> {
dhcp_query()?; dhcp_query()?;
// Init poll timer function
let next_time = next_n_ms_timer_jiffies(5);
let timer = Timer::new(Box::new(NetWorkPollFunc()), next_time);
timer.activate();
return Ok(()); return Ok(());
} }
fn dhcp_query() -> Result<(), SystemError> { fn dhcp_query() -> Result<(), SystemError> {
let binding = NET_DRIVERS.write(); let binding = NET_DRIVERS.write();
let net_face = binding.get(&0).unwrap().clone(); let net_face = binding.get(&0).ok_or(SystemError::ENODEV)?.clone();
drop(binding); drop(binding);
@ -85,3 +112,49 @@ fn dhcp_query() -> Result<(), SystemError> {
return Err(SystemError::ETIMEDOUT); return Err(SystemError::ETIMEDOUT);
} }
pub fn poll_ifaces() {
let guard: RwLockReadGuard<BTreeMap<usize, Arc<dyn NetDriver>>> = NET_DRIVERS.read();
if guard.len() == 0 {
kwarn!("poll_ifaces: No net driver found!");
return;
}
let mut sockets = SOCKET_SET.lock();
for (_, iface) in guard.iter() {
iface.poll(&mut sockets).ok();
}
SOCKET_WAITQUEUE.wakeup_all((-1i64) as u64);
}
/// 对ifaces进行轮询最多对SOCKET_SET尝试times次加锁。
///
/// @return 轮询成功返回Ok(())
/// @return 加锁超时返回SystemError::EAGAIN_OR_EWOULDBLOCK
/// @return 没有网卡返回SystemError::ENODEV
pub fn poll_ifaces_try_lock(times: u16) -> Result<(), SystemError> {
let mut i = 0;
while i < times {
let guard: RwLockReadGuard<BTreeMap<usize, Arc<dyn NetDriver>>> = NET_DRIVERS.read();
if guard.len() == 0 {
kwarn!("poll_ifaces: No net driver found!");
// 没有网卡,返回错误
return Err(SystemError::ENODEV);
}
let sockets = SOCKET_SET.try_lock();
// 加锁失败,继续尝试
if sockets.is_err() {
i += 1;
continue;
}
let mut sockets = sockets.unwrap();
for (_, iface) in guard.iter() {
iface.poll(&mut sockets).ok();
}
SOCKET_WAITQUEUE.wakeup_all((-1i64) as u64);
return Ok(());
}
// 尝试次数用完,返回错误
return Err(SystemError::EAGAIN_OR_EWOULDBLOCK);
}

View File

@ -1,23 +1,36 @@
#![allow(dead_code)] #![allow(dead_code)]
use alloc::{boxed::Box, vec::Vec}; use alloc::{boxed::Box, sync::Arc, vec::Vec};
use smoltcp::{ use smoltcp::{
iface::{SocketHandle, SocketSet}, iface::{SocketHandle, SocketSet},
socket::{raw, tcp, udp}, socket::{raw, tcp, udp},
wire::{IpAddress, IpEndpoint, IpProtocol, Ipv4Address, Ipv4Packet, Ipv6Address}, wire,
}; };
use crate::{ use crate::{
arch::rand::rand, kdebug, kerror, kwarn, libs::spinlock::SpinLock, syscall::SystemError, arch::rand::rand,
driver::net::NetDriver,
filesystem::vfs::{FileType, IndexNode, Metadata, PollStatus},
kerror, kwarn,
libs::{
spinlock::{SpinLock, SpinLockGuard},
wait_queue::WaitQueue,
},
syscall::SystemError,
}; };
use super::{Endpoint, Protocol, Socket, NET_DRIVERS}; use super::{net_core::poll_ifaces, Endpoint, Protocol, Socket, NET_DRIVERS};
lazy_static! { lazy_static! {
/// 所有socket的集合 /// 所有socket的集合
/// TODO: 优化这里自己实现SocketSet现在这样的话不管全局有多少个网卡每个时间点都只会有1个进程能够访问socket /// TODO: 优化这里自己实现SocketSet现在这样的话不管全局有多少个网卡每个时间点都只会有1个进程能够访问socket
pub static ref SOCKET_SET: SpinLock<SocketSet<'static >> = SpinLock::new(SocketSet::new(vec![])); pub static ref SOCKET_SET: SpinLock<SocketSet<'static >> = SpinLock::new(SocketSet::new(vec![]));
pub static ref SOCKET_WAITQUEUE: WaitQueue = WaitQueue::INIT;
} }
/* For setsockopt(2) */
// See: linux-5.19.10/include/uapi/asm-generic/socket.h#9
pub const SOL_SOCKET: u8 = 1;
/// @brief socket的句柄管理组件。 /// @brief socket的句柄管理组件。
/// 它在smoltcp的SocketHandle上封装了一层增加更多的功能。 /// 它在smoltcp的SocketHandle上封装了一层增加更多的功能。
/// 比如在socket被关闭时自动释放socket的资源通知系统的其他组件。 /// 比如在socket被关闭时自动释放socket的资源通知系统的其他组件。
@ -41,6 +54,7 @@ impl Drop for GlobalSocketHandle {
let mut socket_set_guard = SOCKET_SET.lock(); let mut socket_set_guard = SOCKET_SET.lock();
socket_set_guard.remove(self.0); // 删除的时候会发送一条FINISH的信息 socket_set_guard.remove(self.0); // 删除的时候会发送一条FINISH的信息
drop(socket_set_guard); drop(socket_set_guard);
poll_ifaces();
} }
} }
@ -76,15 +90,15 @@ bitflags! {
/// @brief 在trait Socket的metadata函数中返回该结构体供外部使用 /// @brief 在trait Socket的metadata函数中返回该结构体供外部使用
pub struct SocketMetadata { pub struct SocketMetadata {
/// socket的类型 /// socket的类型
socket_type: SocketType, pub socket_type: SocketType,
/// 发送缓冲区的大小 /// 发送缓冲区的大小
send_buf_size: usize, pub send_buf_size: usize,
/// 接收缓冲区的大小 /// 接收缓冲区的大小
recv_buf_size: usize, pub recv_buf_size: usize,
/// 元数据的缓冲区的大小 /// 元数据的缓冲区的大小
metadata_buf_size: usize, pub metadata_buf_size: usize,
/// socket的选项 /// socket的选项
options: SocketOptions, pub options: SocketOptions,
} }
/// @brief 表示原始的socket。原始套接字绕过传输层协议如 TCP 或 UDP并提供对网络层协议如 IP的直接访问。 /// @brief 表示原始的socket。原始套接字绕过传输层协议如 TCP 或 UDP并提供对网络层协议如 IP的直接访问。
@ -127,7 +141,7 @@ impl RawSocket {
let protocol: u8 = protocol.into(); let protocol: u8 = protocol.into();
let socket = raw::Socket::new( let socket = raw::Socket::new(
smoltcp::wire::IpVersion::Ipv4, smoltcp::wire::IpVersion::Ipv4,
IpProtocol::from(protocol), wire::IpProtocol::from(protocol),
tx_buffer, tx_buffer,
rx_buffer, rx_buffer,
); );
@ -144,7 +158,8 @@ impl RawSocket {
} }
impl Socket for RawSocket { impl Socket for RawSocket {
fn read(&self, buf: &mut [u8]) -> Result<(usize, Endpoint), SystemError> { fn read(&self, buf: &mut [u8]) -> (Result<usize, SystemError>, Endpoint) {
poll_ifaces();
loop { loop {
// 如何优化这里? // 如何优化这里?
let mut socket_set_guard = SOCKET_SET.lock(); let mut socket_set_guard = SOCKET_SET.lock();
@ -152,24 +167,25 @@ impl Socket for RawSocket {
match socket.recv_slice(buf) { match socket.recv_slice(buf) {
Ok(len) => { Ok(len) => {
let packet = Ipv4Packet::new_unchecked(buf); let packet = wire::Ipv4Packet::new_unchecked(buf);
return Ok(( return (
len, Ok(len),
Endpoint::Ip(smoltcp::wire::IpEndpoint { Endpoint::Ip(Some(smoltcp::wire::IpEndpoint {
addr: IpAddress::Ipv4(packet.src_addr()), addr: wire::IpAddress::Ipv4(packet.src_addr()),
port: 0, port: 0,
}), })),
)); );
} }
Err(smoltcp::socket::raw::RecvError::Exhausted) => { Err(smoltcp::socket::raw::RecvError::Exhausted) => {
if !self.options.contains(SocketOptions::BLOCK) { if !self.options.contains(SocketOptions::BLOCK) {
// 如果是非阻塞的socket就返回错误 // 如果是非阻塞的socket就返回错误
return Err(SystemError::EAGAIN_OR_EWOULDBLOCK); return (Err(SystemError::EAGAIN_OR_EWOULDBLOCK), Endpoint::Ip(None));
} }
} }
} }
drop(socket); drop(socket);
drop(socket_set_guard); drop(socket_set_guard);
SOCKET_WAITQUEUE.sleep();
} }
} }
@ -189,7 +205,7 @@ impl Socket for RawSocket {
} else { } else {
// 如果用户发送的数据包不包含IP头则需要自己构造IP头 // 如果用户发送的数据包不包含IP头则需要自己构造IP头
if let Some(Endpoint::Ip(endpoint)) = to { if let Some(Endpoint::Ip(Some(endpoint))) = to {
let mut socket_set_guard = SOCKET_SET.lock(); let mut socket_set_guard = SOCKET_SET.lock();
let socket: &mut raw::Socket = let socket: &mut raw::Socket =
socket_set_guard.get_mut::<raw::Socket>(self.handle.0); socket_set_guard.get_mut::<raw::Socket>(self.handle.0);
@ -205,13 +221,13 @@ impl Socket for RawSocket {
} }
let ipv4_src_addr = ipv4_src_addr.unwrap(); let ipv4_src_addr = ipv4_src_addr.unwrap();
if let IpAddress::Ipv4(ipv4_dst) = endpoint.addr { if let wire::IpAddress::Ipv4(ipv4_dst) = endpoint.addr {
let len = buf.len(); let len = buf.len();
// 创建20字节的IPv4头部 // 创建20字节的IPv4头部
let mut buffer: Vec<u8> = vec![0u8; len + 20]; let mut buffer: Vec<u8> = vec![0u8; len + 20];
let mut packet: Ipv4Packet<&mut Vec<u8>> = let mut packet: wire::Ipv4Packet<&mut Vec<u8>> =
Ipv4Packet::new_unchecked(&mut buffer); wire::Ipv4Packet::new_unchecked(&mut buffer);
// 封装ipv4 header // 封装ipv4 header
packet.set_version(4); packet.set_version(4);
@ -234,9 +250,10 @@ impl Socket for RawSocket {
socket.send_slice(&buffer).unwrap(); socket.send_slice(&buffer).unwrap();
drop(socket); drop(socket);
drop(socket_set_guard);
// poll? iface.poll(&mut socket_set_guard).ok();
drop(socket_set_guard);
return Ok(len); return Ok(len);
} else { } else {
kwarn!("Unsupport Ip protocol type!"); kwarn!("Unsupport Ip protocol type!");
@ -306,113 +323,143 @@ impl UdpSocket {
options, options,
}; };
} }
fn do_bind(&self, socket: &mut udp::Socket, endpoint: Endpoint) -> Result<(), SystemError> {
if let Endpoint::Ip(Some(ip)) = endpoint {
let bind_res = if ip.addr.is_unspecified() {
socket.bind(ip.port)
} else {
socket.bind(ip)
};
match bind_res {
Ok(()) => return Ok(()),
Err(_) => return Err(SystemError::EINVAL),
}
} else {
return Err(SystemError::EINVAL);
};
}
} }
impl Socket for UdpSocket { impl Socket for UdpSocket {
/// @brief 在read函数执行之前请先bind到本地的指定端口 /// @brief 在read函数执行之前请先bind到本地的指定端口
fn read(&self, buf: &mut [u8]) -> Result<(usize, Endpoint), SystemError> { fn read(&self, buf: &mut [u8]) -> (Result<usize, SystemError>, Endpoint) {
loop { loop {
kdebug!("Wait22 to Read"); // kdebug!("Wait22 to Read");
poll_ifaces();
let mut socket_set_guard = SOCKET_SET.lock(); let mut socket_set_guard = SOCKET_SET.lock();
let socket = socket_set_guard.get_mut::<udp::Socket>(self.handle.0); let socket = socket_set_guard.get_mut::<udp::Socket>(self.handle.0);
kdebug!("Wait to Read"); // kdebug!("Wait to Read");
if socket.can_recv() { if socket.can_recv() {
kdebug!("Can Receive"); if let Ok((size, remote_endpoint)) = socket.recv_slice(buf) {
if let Ok((size, endpoint)) = socket.recv_slice(buf) {
drop(socket); drop(socket);
drop(socket_set_guard); drop(socket_set_guard);
poll_ifaces();
return Ok((size, Endpoint::Ip(endpoint))); return (Ok(size), Endpoint::Ip(Some(remote_endpoint)));
} }
} else { } else {
// 没有数据可以读取. 如果没有bind到指定端口也会导致rx_buf为空 // 如果socket没有连接则忙等
return Err(SystemError::EAGAIN_OR_EWOULDBLOCK); // return (Err(SystemError::ENOTCONN), Endpoint::Ip(None));
} }
drop(socket);
drop(socket_set_guard);
SOCKET_WAITQUEUE.sleep();
} }
} }
fn write(&self, buf: &[u8], to: Option<super::Endpoint>) -> Result<usize, SystemError> { fn write(&self, buf: &[u8], to: Option<super::Endpoint>) -> Result<usize, SystemError> {
let endpoint: &IpEndpoint = { // kdebug!("udp to send: {:?}, len={}", to, buf.len());
if let Some(Endpoint::Ip(ref endpoint)) = to { let remote_endpoint: &wire::IpEndpoint = {
if let Some(Endpoint::Ip(Some(ref endpoint))) = to {
endpoint endpoint
} else if let Some(Endpoint::Ip(ref endpoint)) = self.remote_endpoint { } else if let Some(Endpoint::Ip(Some(ref endpoint))) = self.remote_endpoint {
endpoint endpoint
} else { } else {
return Err(SystemError::ENOTCONN); return Err(SystemError::ENOTCONN);
} }
}; };
// kdebug!("udp write: remote = {:?}", remote_endpoint);
let mut socket_set_guard = SOCKET_SET.lock(); let mut socket_set_guard = SOCKET_SET.lock();
let socket = socket_set_guard.get_mut::<udp::Socket>(self.handle.0); let socket = socket_set_guard.get_mut::<udp::Socket>(self.handle.0);
// kdebug!("is open()={}", socket.is_open());
// kdebug!("socket endpoint={:?}", socket.endpoint());
if socket.endpoint().port == 0 { if socket.endpoint().port == 0 {
let temp_port = get_ephemeral_port(); let temp_port = get_ephemeral_port();
match endpoint.addr { let local_ep = match remote_endpoint.addr {
// 远程remote endpoint使用什么协议发送的时候使用的协议是一样的吧 // 远程remote endpoint使用什么协议发送的时候使用的协议是一样的吧
// 否则就用 self.endpoint().addr.unwrap() // 否则就用 self.endpoint().addr.unwrap()
IpAddress::Ipv4(_) => { wire::IpAddress::Ipv4(_) => Endpoint::Ip(Some(wire::IpEndpoint::new(
socket smoltcp::wire::IpAddress::Ipv4(wire::Ipv4Address::UNSPECIFIED),
.bind(IpEndpoint::new( temp_port,
smoltcp::wire::IpAddress::Ipv4(Ipv4Address::UNSPECIFIED), ))),
temp_port, wire::IpAddress::Ipv6(_) => Endpoint::Ip(Some(wire::IpEndpoint::new(
)) smoltcp::wire::IpAddress::Ipv6(wire::Ipv6Address::UNSPECIFIED),
.unwrap(); temp_port,
} ))),
IpAddress::Ipv6(_) => { };
socket // kdebug!("udp write: local_ep = {:?}", local_ep);
.bind(IpEndpoint::new( self.do_bind(socket, local_ep)?;
smoltcp::wire::IpAddress::Ipv6(Ipv6Address::UNSPECIFIED),
temp_port,
))
.unwrap();
}
}
} }
// kdebug!("is open()={}", socket.is_open());
return if socket.can_send() { if socket.can_send() {
match socket.send_slice(&buf, *endpoint) { // kdebug!("udp write: can send");
match socket.send_slice(&buf, *remote_endpoint) {
Ok(()) => { Ok(()) => {
// avoid deadlock // kdebug!("udp write: send ok");
drop(socket); drop(socket);
drop(socket_set_guard); drop(socket_set_guard);
poll_ifaces();
Ok(buf.len()) return Ok(buf.len());
}
Err(_) => {
// kdebug!("udp write: send err");
return Err(SystemError::ENOBUFS);
} }
Err(_) => Err(SystemError::ENOBUFS),
} }
} else { } else {
Err(SystemError::ENOBUFS) // kdebug!("udp write: can not send");
return Err(SystemError::ENOBUFS);
}; };
} }
fn bind(&self, endpoint: Endpoint) -> Result<(), SystemError> { fn bind(&mut self, endpoint: Endpoint) -> Result<(), SystemError> {
let mut sockets = SOCKET_SET.lock(); let mut sockets = SOCKET_SET.lock();
let socket = sockets.get_mut::<udp::Socket>(self.handle.0); let socket = sockets.get_mut::<udp::Socket>(self.handle.0);
// kdebug!("UDP Bind to {:?}", endpoint);
return self.do_bind(socket, endpoint);
}
return if let Endpoint::Ip(ip) = endpoint { fn poll(&self) -> (bool, bool, bool) {
match socket.bind(ip) { let sockets = SOCKET_SET.lock();
Ok(()) => Ok(()), let socket = sockets.get::<udp::Socket>(self.handle.0);
Err(_) => Err(SystemError::EINVAL),
} return (socket.can_send(), socket.can_recv(), false);
} else {
Err(SystemError::EINVAL)
};
} }
/// @brief /// @brief
fn connect(&mut self, endpoint: super::Endpoint) -> Result<(), SystemError> { fn connect(&mut self, endpoint: super::Endpoint) -> Result<(), SystemError> {
return if let Endpoint::Ip(_) = endpoint { if let Endpoint::Ip(_) = endpoint {
self.remote_endpoint = Some(endpoint); self.remote_endpoint = Some(endpoint);
Ok(()) return Ok(());
} else { } else {
Err(SystemError::EINVAL) return Err(SystemError::EINVAL);
}; };
} }
fn ioctl(
&self,
_cmd: usize,
_arg0: usize,
_arg1: usize,
_arg2: usize,
) -> Result<usize, SystemError> {
todo!()
}
fn metadata(&self) -> Result<SocketMetadata, SystemError> { fn metadata(&self) -> Result<SocketMetadata, SystemError> {
todo!() todo!()
} }
@ -420,6 +467,31 @@ impl Socket for UdpSocket {
fn box_clone(&self) -> alloc::boxed::Box<dyn Socket> { fn box_clone(&self) -> alloc::boxed::Box<dyn Socket> {
return Box::new(self.clone()); return Box::new(self.clone());
} }
fn endpoint(&self) -> Option<Endpoint> {
let sockets = SOCKET_SET.lock();
let socket = sockets.get::<udp::Socket>(self.handle.0);
let listen_endpoint = socket.endpoint();
if listen_endpoint.port == 0 {
return None;
} else {
// 如果listen_endpoint的address是None意味着“监听所有的地址”。
// 这里假设所有的地址都是ipv4
// TODO: 支持ipv6
let result = wire::IpEndpoint::new(
listen_endpoint
.addr
.unwrap_or(wire::IpAddress::v4(0, 0, 0, 0)),
listen_endpoint.port,
);
return Some(Endpoint::Ip(Some(result)));
}
}
fn peer_endpoint(&self) -> Option<Endpoint> {
return self.remote_endpoint.clone();
}
} }
/// @brief 表示 tcp socket /// @brief 表示 tcp socket
@ -428,7 +500,7 @@ impl Socket for UdpSocket {
#[derive(Debug, Clone)] #[derive(Debug, Clone)]
pub struct TcpSocket { pub struct TcpSocket {
handle: GlobalSocketHandle, handle: GlobalSocketHandle,
local_endpoint: Option<IpEndpoint>, // save local endpoint for bind() local_endpoint: Option<wire::IpEndpoint>, // save local endpoint for bind()
is_listening: bool, is_listening: bool,
options: SocketOptions, options: SocketOptions,
} }
@ -437,9 +509,9 @@ impl TcpSocket {
/// 元数据的缓冲区的大小 /// 元数据的缓冲区的大小
pub const DEFAULT_METADATA_BUF_SIZE: usize = 1024; pub const DEFAULT_METADATA_BUF_SIZE: usize = 1024;
/// 默认的发送缓冲区的大小 transmiss /// 默认的发送缓冲区的大小 transmiss
pub const DEFAULT_RX_BUF_SIZE: usize = 64 * 1024; pub const DEFAULT_RX_BUF_SIZE: usize = 512 * 1024;
/// 默认的接收缓冲区的大小 receive /// 默认的接收缓冲区的大小 receive
pub const DEFAULT_TX_BUF_SIZE: usize = 64 * 1024; pub const DEFAULT_TX_BUF_SIZE: usize = 512 * 1024;
/// @brief 创建一个原始的socket /// @brief 创建一个原始的socket
/// ///
@ -462,33 +534,83 @@ impl TcpSocket {
options, options,
}; };
} }
fn do_listen(
&mut self,
socket: &mut smoltcp::socket::tcp::Socket,
local_endpoint: smoltcp::wire::IpEndpoint,
) -> Result<(), SystemError> {
let listen_result = if local_endpoint.addr.is_unspecified() {
// kdebug!("Tcp Socket Listen on port {}", local_endpoint.port);
socket.listen(local_endpoint.port)
} else {
// kdebug!("Tcp Socket Listen on {local_endpoint}");
socket.listen(local_endpoint)
};
// todo: 增加端口占用检查
return match listen_result {
Ok(()) => {
// kdebug!(
// "Tcp Socket Listen on {local_endpoint}, open?:{}",
// socket.is_open()
// );
self.is_listening = true;
Ok(())
}
Err(_) => Err(SystemError::EINVAL),
};
}
} }
impl Socket for TcpSocket { impl Socket for TcpSocket {
/// @breif fn read(&self, buf: &mut [u8]) -> (Result<usize, SystemError>, Endpoint) {
fn read(&self, buf: &mut [u8]) -> Result<(usize, Endpoint), SystemError> { // kdebug!("tcp socket: read, buf len={}", buf.len());
loop { loop {
poll_ifaces();
let mut socket_set_guard = SOCKET_SET.lock(); let mut socket_set_guard = SOCKET_SET.lock();
let socket = socket_set_guard.get_mut::<tcp::Socket>(self.handle.0); let socket = socket_set_guard.get_mut::<tcp::Socket>(self.handle.0);
// 如果socket已经关闭返回错误
if !socket.is_active() {
// kdebug!("Tcp Socket Read Error, socket is closed");
return (Err(SystemError::ENOTCONN), Endpoint::Ip(None));
}
if socket.may_recv() { if socket.may_recv() {
if let Ok(size) = socket.recv_slice(buf) { let recv_res = socket.recv_slice(buf);
if let Ok(size) = recv_res {
if size > 0 { if size > 0 {
let endpoint = if let Some(p) = socket.remote_endpoint() { let endpoint = if let Some(p) = socket.remote_endpoint() {
p p
} else { } else {
return Err(SystemError::ENOTCONN); return (Err(SystemError::ENOTCONN), Endpoint::Ip(None));
}; };
drop(socket); drop(socket);
drop(socket_set_guard); drop(socket_set_guard);
poll_ifaces();
return Ok((size, Endpoint::Ip(endpoint))); return (Ok(size), Endpoint::Ip(Some(endpoint)));
}
} else {
let err = recv_res.unwrap_err();
match err {
tcp::RecvError::InvalidState => {
kwarn!("Tcp Socket Read Error, InvalidState");
return (Err(SystemError::ENOTCONN), Endpoint::Ip(None));
}
tcp::RecvError::Finished => {
return (Err(SystemError::ENOTCONN), Endpoint::Ip(None));
}
} }
} }
} else { } else {
return Err(SystemError::EAGAIN_OR_EWOULDBLOCK); return (Err(SystemError::ENOTCONN), Endpoint::Ip(None));
} }
drop(socket);
drop(socket_set_guard);
SOCKET_WAITQUEUE.sleep();
} }
} }
@ -502,7 +624,7 @@ impl Socket for TcpSocket {
Ok(size) => { Ok(size) => {
drop(socket); drop(socket);
drop(socket_set_guard); drop(socket_set_guard);
poll_ifaces();
return Ok(size); return Ok(size);
} }
Err(e) => { Err(e) => {
@ -518,54 +640,83 @@ impl Socket for TcpSocket {
return Err(SystemError::ENOTCONN); return Err(SystemError::ENOTCONN);
} }
fn connect(&mut self, _endpoint: super::Endpoint) -> Result<(), SystemError> { fn poll(&self) -> (bool, bool, bool) {
// let mut sockets = SOCKET_SET.lock(); let mut socket_set_guard = SOCKET_SET.lock();
// let mut socket = sockets.get::<tcp::Socket>(self.handle.0); let socket = socket_set_guard.get_mut::<tcp::Socket>(self.handle.0);
// if let Endpoint::Ip(ip) = endpoint { let mut input = false;
// let temp_port = if ip.port == 0 { let mut output = false;
// get_ephemeral_port() let mut error = false;
// } else { if self.is_listening && socket.is_active() {
// ip.port input = true;
// }; } else if !socket.is_open() {
error = true;
} else {
if socket.may_recv() {
input = true;
}
if socket.can_send() {
output = true;
}
}
// return match socket.connect(iface.context(), temp_port) { return (input, output, error);
// Ok(()) => { }
// // avoid deadlock
// drop(socket);
// drop(sockets);
// // wait for connection result fn connect(&mut self, endpoint: Endpoint) -> Result<(), SystemError> {
// loop { let mut sockets = SOCKET_SET.lock();
// // poll_ifaces(); let socket = sockets.get_mut::<tcp::Socket>(self.handle.0);
// let mut sockets = SOCKET_SET.lock(); if let Endpoint::Ip(Some(ip)) = endpoint {
// let socket = sockets.get::<tcp::Socket>(self.handle.0); let temp_port = get_ephemeral_port();
// match socket.state() { // kdebug!("temp_port: {}", temp_port);
// State::SynSent => { let iface: Arc<dyn NetDriver> = NET_DRIVERS.write().get(&0).unwrap().clone();
// // still connecting let mut inner_iface = iface.inner_iface().lock();
// drop(socket); // kdebug!("to connect: {ip:?}");
// kdebug!("poll for connection wait");
// // SOCKET_ACTIVITY.wait(sockets); match socket.connect(&mut inner_iface.context(), ip, temp_port) {
// } Ok(()) => {
// State::Established => { // avoid deadlock
// break Ok(()); drop(inner_iface);
// } drop(iface);
// _ => { drop(socket);
// break Err(SystemError::ECONNREFUSED); drop(sockets);
// } loop {
// } poll_ifaces();
// } let mut sockets = SOCKET_SET.lock();
// } let socket = sockets.get_mut::<tcp::Socket>(self.handle.0);
// Err(_) => Err(SystemError::ENOBUFS),
// }; match socket.state() {
// } else { tcp::State::Established => {
// return Err(SystemError::EINVAL); return Ok(());
// } }
return Err(SystemError::EINVAL); tcp::State::SynSent => {
drop(socket);
drop(sockets);
SOCKET_WAITQUEUE.sleep();
}
_ => {
return Err(SystemError::ECONNREFUSED);
}
}
}
}
Err(e) => {
// kerror!("Tcp Socket Connect Error {e:?}");
match e {
tcp::ConnectError::InvalidState => return Err(SystemError::EISCONN),
tcp::ConnectError::Unaddressable => return Err(SystemError::EADDRNOTAVAIL),
}
}
}
} else {
return Err(SystemError::EINVAL);
}
} }
/// @brief tcp socket 监听 local_endpoint 端口 /// @brief tcp socket 监听 local_endpoint 端口
///
/// @param backlog 未处理的连接队列的最大长度. 由于smoltcp不支持backlog所以这个参数目前无效
fn listen(&mut self, _backlog: usize) -> Result<(), SystemError> { fn listen(&mut self, _backlog: usize) -> Result<(), SystemError> {
if self.is_listening { if self.is_listening {
return Ok(()); return Ok(());
@ -576,16 +727,101 @@ impl Socket for TcpSocket {
let socket = sockets.get_mut::<tcp::Socket>(self.handle.0); let socket = sockets.get_mut::<tcp::Socket>(self.handle.0);
if socket.is_listening() { if socket.is_listening() {
// kdebug!("Tcp Socket is already listening on {local_endpoint}");
return Ok(()); return Ok(());
} }
// kdebug!("Tcp Socket before listen, open={}", socket.is_open());
return self.do_listen(socket, local_endpoint);
}
return match socket.listen(local_endpoint) { fn bind(&mut self, endpoint: Endpoint) -> Result<(), SystemError> {
Ok(()) => { if let Endpoint::Ip(Some(mut ip)) = endpoint {
self.is_listening = true; if ip.port == 0 {
Ok(()) ip.port = get_ephemeral_port();
} }
Err(_) => Err(SystemError::EINVAL),
}; self.local_endpoint = Some(ip);
self.is_listening = false;
return Ok(());
}
return Err(SystemError::EINVAL);
}
fn shutdown(&self, _type: super::ShutdownType) -> Result<(), SystemError> {
let mut sockets = SOCKET_SET.lock();
let socket = sockets.get_mut::<tcp::Socket>(self.handle.0);
socket.close();
return Ok(());
}
fn accept(&mut self) -> Result<(Box<dyn Socket>, Endpoint), SystemError> {
let endpoint = self.local_endpoint.ok_or(SystemError::EINVAL)?;
loop {
// kdebug!("tcp accept: poll_ifaces()");
poll_ifaces();
let mut sockets = SOCKET_SET.lock();
let socket = sockets.get_mut::<tcp::Socket>(self.handle.0);
if socket.is_active() {
// kdebug!("tcp accept: socket.is_active()");
let remote_ep = socket.remote_endpoint().ok_or(SystemError::ENOTCONN)?;
drop(socket);
let new_socket = {
// Initialize the TCP socket's buffers.
let rx_buffer = tcp::SocketBuffer::new(vec![0; Self::DEFAULT_RX_BUF_SIZE]);
let tx_buffer = tcp::SocketBuffer::new(vec![0; Self::DEFAULT_TX_BUF_SIZE]);
// The new TCP socket used for sending and receiving data.
let mut tcp_socket = tcp::Socket::new(rx_buffer, tx_buffer);
self.do_listen(&mut tcp_socket, endpoint)
.expect("do_listen failed");
// tcp_socket.listen(endpoint).unwrap();
// 之所以把old_handle存入new_socket, 是因为当前时刻smoltcp已经把old_handle对应的socket与远程的endpoint关联起来了
// 因此需要再为当前的socket分配一个新的handle
let new_handle = GlobalSocketHandle::new(sockets.add(tcp_socket));
let old_handle = ::core::mem::replace(&mut self.handle, new_handle);
Box::new(TcpSocket {
handle: old_handle,
local_endpoint: self.local_endpoint,
is_listening: false,
options: self.options,
})
};
// kdebug!("tcp accept: new socket: {:?}", new_socket);
drop(sockets);
poll_ifaces();
return Ok((new_socket, Endpoint::Ip(Some(remote_ep))));
}
drop(socket);
drop(sockets);
SOCKET_WAITQUEUE.sleep();
}
}
fn endpoint(&self) -> Option<Endpoint> {
let mut result: Option<Endpoint> =
self.local_endpoint.clone().map(|x| Endpoint::Ip(Some(x)));
if result.is_none() {
let sockets = SOCKET_SET.lock();
let socket = sockets.get::<tcp::Socket>(self.handle.0);
if let Some(ep) = socket.local_endpoint() {
result = Some(Endpoint::Ip(Some(ep)));
}
}
return result;
}
fn peer_endpoint(&self) -> Option<Endpoint> {
let sockets = SOCKET_SET.lock();
let socket = sockets.get::<tcp::Socket>(self.handle.0);
return socket.remote_endpoint().map(|x| Endpoint::Ip(Some(x)));
} }
fn metadata(&self) -> Result<SocketMetadata, SystemError> { fn metadata(&self) -> Result<SocketMetadata, SystemError> {
@ -598,8 +834,11 @@ impl Socket for TcpSocket {
} }
/// @breif 自动分配一个未被使用的PORT /// @breif 自动分配一个未被使用的PORT
///
/// TODO: 增加ListenTable, 用于检查端口是否被占用
pub fn get_ephemeral_port() -> u16 { pub fn get_ephemeral_port() -> u16 {
// TODO selects non-conflict high port // TODO selects non-conflict high port
static mut EPHEMERAL_PORT: u16 = 0; static mut EPHEMERAL_PORT: u16 = 0;
unsafe { unsafe {
if EPHEMERAL_PORT == 0 { if EPHEMERAL_PORT == 0 {
@ -613,3 +852,223 @@ pub fn get_ephemeral_port() -> u16 {
EPHEMERAL_PORT EPHEMERAL_PORT
} }
} }
/// @brief 地址族的枚举
///
/// 参考https://opengrok.ringotek.cn/xref/linux-5.19.10/include/linux/socket.h#180
#[derive(Debug, Clone, Copy, PartialEq, Eq, FromPrimitive, ToPrimitive)]
pub enum AddressFamily {
/// AF_UNSPEC 表示地址族未指定
Unspecified = 0,
/// AF_UNIX 表示Unix域的socket (与AF_LOCAL相同)
Unix = 1,
/// AF_INET 表示IPv4的socket
INet = 2,
/// AF_AX25 表示AMPR AX.25的socket
AX25 = 3,
/// AF_IPX 表示IPX的socket
IPX = 4,
/// AF_APPLETALK 表示Appletalk的socket
Appletalk = 5,
/// AF_NETROM 表示AMPR NET/ROM的socket
Netrom = 6,
/// AF_BRIDGE 表示多协议桥接的socket
Bridge = 7,
/// AF_ATMPVC 表示ATM PVCs的socket
Atmpvc = 8,
/// AF_X25 表示X.25的socket
X25 = 9,
/// AF_INET6 表示IPv6的socket
INet6 = 10,
/// AF_ROSE 表示AMPR ROSE的socket
Rose = 11,
/// AF_DECnet Reserved for DECnet project
Decnet = 12,
/// AF_NETBEUI Reserved for 802.2LLC project
Netbeui = 13,
/// AF_SECURITY 表示Security callback的伪AF
Security = 14,
/// AF_KEY 表示Key management API
Key = 15,
/// AF_NETLINK 表示Netlink的socket
Netlink = 16,
/// AF_PACKET 表示Low level packet interface
Packet = 17,
/// AF_ASH 表示Ash
Ash = 18,
/// AF_ECONET 表示Acorn Econet
Econet = 19,
/// AF_ATMSVC 表示ATM SVCs
Atmsvc = 20,
/// AF_RDS 表示Reliable Datagram Sockets
Rds = 21,
/// AF_SNA 表示Linux SNA Project
Sna = 22,
/// AF_IRDA 表示IRDA sockets
Irda = 23,
/// AF_PPPOX 表示PPPoX sockets
Pppox = 24,
/// AF_WANPIPE 表示WANPIPE API sockets
WanPipe = 25,
/// AF_LLC 表示Linux LLC
Llc = 26,
/// AF_IB 表示Native InfiniBand address
/// 介绍https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/9/html-single/configuring_infiniband_and_rdma_networks/index#understanding-infiniband-and-rdma_configuring-infiniband-and-rdma-networks
Ib = 27,
/// AF_MPLS 表示MPLS
Mpls = 28,
/// AF_CAN 表示Controller Area Network
Can = 29,
/// AF_TIPC 表示TIPC sockets
Tipc = 30,
/// AF_BLUETOOTH 表示Bluetooth sockets
Bluetooth = 31,
/// AF_IUCV 表示IUCV sockets
Iucv = 32,
/// AF_RXRPC 表示RxRPC sockets
Rxrpc = 33,
/// AF_ISDN 表示mISDN sockets
Isdn = 34,
/// AF_PHONET 表示Phonet sockets
Phonet = 35,
/// AF_IEEE802154 表示IEEE 802.15.4 sockets
Ieee802154 = 36,
/// AF_CAIF 表示CAIF sockets
Caif = 37,
/// AF_ALG 表示Algorithm sockets
Alg = 38,
/// AF_NFC 表示NFC sockets
Nfc = 39,
/// AF_VSOCK 表示vSockets
Vsock = 40,
/// AF_KCM 表示Kernel Connection Multiplexor
Kcm = 41,
/// AF_QIPCRTR 表示Qualcomm IPC Router
Qipcrtr = 42,
/// AF_SMC 表示SMC-R sockets.
/// reserve number for PF_SMC protocol family that reuses AF_INET address family
Smc = 43,
/// AF_XDP 表示XDP sockets
Xdp = 44,
/// AF_MCTP 表示Management Component Transport Protocol
Mctp = 45,
/// AF_MAX 表示最大的地址族
Max = 46,
}
impl TryFrom<u16> for AddressFamily {
type Error = SystemError;
fn try_from(x: u16) -> Result<Self, Self::Error> {
use num_traits::FromPrimitive;
return <Self as FromPrimitive>::from_u16(x).ok_or_else(|| SystemError::EINVAL);
}
}
/// @brief posix套接字类型的枚举(这些值与linux内核中的值一致)
#[derive(Debug, Clone, Copy, PartialEq, Eq, FromPrimitive, ToPrimitive)]
pub enum PosixSocketType {
Stream = 1,
Datagram = 2,
Raw = 3,
Rdm = 4,
SeqPacket = 5,
Dccp = 6,
Packet = 10,
}
impl TryFrom<u8> for PosixSocketType {
type Error = SystemError;
fn try_from(x: u8) -> Result<Self, Self::Error> {
use num_traits::FromPrimitive;
return <Self as FromPrimitive>::from_u8(x).ok_or_else(|| SystemError::EINVAL);
}
}
/// @brief Socket在文件系统中的inode封装
#[derive(Debug)]
pub struct SocketInode(SpinLock<Box<dyn Socket>>);
impl SocketInode {
pub fn new(socket: Box<dyn Socket>) -> Arc<Self> {
return Arc::new(Self(SpinLock::new(socket)));
}
#[inline]
pub fn inner(&self) -> SpinLockGuard<Box<dyn Socket>> {
return self.0.lock();
}
}
impl IndexNode for SocketInode {
fn open(
&self,
_data: &mut crate::filesystem::vfs::FilePrivateData,
_mode: &crate::filesystem::vfs::file::FileMode,
) -> Result<(), SystemError> {
return Ok(());
}
fn close(
&self,
_data: &mut crate::filesystem::vfs::FilePrivateData,
) -> Result<(), SystemError> {
return Ok(());
}
fn read_at(
&self,
_offset: usize,
len: usize,
buf: &mut [u8],
_data: &mut crate::filesystem::vfs::FilePrivateData,
) -> Result<usize, SystemError> {
return self.0.lock().read(&mut buf[0..len]).0;
}
fn write_at(
&self,
_offset: usize,
len: usize,
buf: &[u8],
_data: &mut crate::filesystem::vfs::FilePrivateData,
) -> Result<usize, SystemError> {
return self.0.lock().write(&buf[0..len], None);
}
fn poll(&self) -> Result<crate::filesystem::vfs::PollStatus, SystemError> {
let (read, write, error) = self.0.lock().poll();
let mut result = PollStatus::empty();
if read {
result.insert(PollStatus::READ);
}
if write {
result.insert(PollStatus::WRITE);
}
if error {
result.insert(PollStatus::ERROR);
}
return Ok(result);
}
fn fs(&self) -> alloc::sync::Arc<dyn crate::filesystem::vfs::FileSystem> {
todo!()
}
fn as_any_ref(&self) -> &dyn core::any::Any {
self
}
fn list(&self) -> Result<Vec<alloc::string::String>, SystemError> {
return Err(SystemError::ENOTDIR);
}
fn metadata(&self) -> Result<crate::filesystem::vfs::Metadata, SystemError> {
let meta = Metadata {
mode: 0o777,
file_type: FileType::Socket,
..Default::default()
};
return Ok(meta);
}
}

1122
kernel/src/net/syscall.rs Normal file

File diff suppressed because it is too large Load Diff

View File

@ -3,18 +3,20 @@ use core::{
ptr::{null_mut, read_volatile, write_volatile}, ptr::{null_mut, read_volatile, write_volatile},
}; };
use alloc::boxed::Box; use alloc::{boxed::Box, sync::Arc};
use crate::{ use crate::{
arch::{asm::current::current_pcb, fpu::FpState}, arch::{asm::current::current_pcb, fpu::FpState},
filesystem::vfs::{ filesystem::vfs::{
file::{File, FileDescriptorVec, FileMode}, file::{File, FileDescriptorVec, FileMode},
ROOT_INODE, FileType, ROOT_INODE,
}, },
include::bindings::bindings::{ include::bindings::bindings::{
process_control_block, CLONE_FS, PROC_INTERRUPTIBLE, PROC_RUNNING, PROC_STOPPED, process_control_block, CLONE_FS, PROC_INTERRUPTIBLE, PROC_RUNNING, PROC_STOPPED,
PROC_UNINTERRUPTIBLE, PROC_UNINTERRUPTIBLE,
}, },
libs::casting::DowncastArc,
net::socket::SocketInode,
sched::core::{cpu_executing, sched_enqueue}, sched::core::{cpu_executing, sched_enqueue},
smp::core::{smp_get_processor_id, smp_send_reschedule}, smp::core::{smp_get_processor_id, smp_send_reschedule},
syscall::SystemError, syscall::SystemError,
@ -291,6 +293,24 @@ impl process_control_block {
pub unsafe fn mark_sleep_uninterruptible(&mut self) { pub unsafe fn mark_sleep_uninterruptible(&mut self) {
self.state = PROC_UNINTERRUPTIBLE as u64; self.state = PROC_UNINTERRUPTIBLE as u64;
} }
/// @brief 根据文件描述符序号获取socket对象的可变引用
///
/// @param fd 文件描述符序号
///
/// @return Option(&mut Box<dyn Socket>) socket对象的可变引用. 如果文件描述符不是socket那么返回None
pub fn get_socket(&self, fd: i32) -> Option<Arc<SocketInode>> {
let f = self.get_file_mut_by_fd(fd)?;
if f.file_type() != FileType::Socket {
return None;
}
let socket: Arc<SocketInode> = f
.inode()
.downcast_arc::<SocketInode>()
.expect("Not a socket inode");
return Some(socket);
}
} }
// =========== 导出到C的函数在将来进程管理模块被完全重构之后需要删掉他们 BEGIN ============ // =========== 导出到C的函数在将来进程管理模块被完全重构之后需要删掉他们 BEGIN ============

View File

@ -22,6 +22,21 @@ extern uint64_t sys_sigaction(struct pt_regs *regs);
extern uint64_t sys_rt_sigreturn(struct pt_regs *regs); extern uint64_t sys_rt_sigreturn(struct pt_regs *regs);
extern uint64_t sys_getpid(struct pt_regs *regs); extern uint64_t sys_getpid(struct pt_regs *regs);
extern uint64_t sys_sched(struct pt_regs *regs); extern uint64_t sys_sched(struct pt_regs *regs);
extern int sys_dup(int oldfd);
extern int sys_dup2(int oldfd, int newfd);
extern uint64_t sys_socket(struct pt_regs *regs);
extern uint64_t sys_setsockopt(struct pt_regs *regs);
extern uint64_t sys_getsockopt(struct pt_regs *regs);
extern uint64_t sys_connect(struct pt_regs *regs);
extern uint64_t sys_bind(struct pt_regs *regs);
extern uint64_t sys_sendto(struct pt_regs *regs);
extern uint64_t sys_recvfrom(struct pt_regs *regs);
extern uint64_t sys_recvmsg(struct pt_regs *regs);
extern uint64_t sys_listen(struct pt_regs *regs);
extern uint64_t sys_shutdown(struct pt_regs *regs);
extern uint64_t sys_accept(struct pt_regs *regs);
extern uint64_t sys_getsockname(struct pt_regs *regs);
extern uint64_t sys_getpeername(struct pt_regs *regs);
/** /**
* @brief * @brief
@ -179,9 +194,9 @@ uint64_t sys_brk(struct pt_regs *regs)
// kdebug("sys_brk input= %#010lx , new_brk= %#010lx bytes current_pcb->mm->brk_start=%#018lx // kdebug("sys_brk input= %#010lx , new_brk= %#010lx bytes current_pcb->mm->brk_start=%#018lx
// current->end_brk=%#018lx", regs->r8, new_brk, current_pcb->mm->brk_start, current_pcb->mm->brk_end); // current->end_brk=%#018lx", regs->r8, new_brk, current_pcb->mm->brk_start, current_pcb->mm->brk_end);
struct mm_struct *mm = current_pcb->mm; struct mm_struct *mm = current_pcb->mm;
if (new_brk < mm->brk_start || new_brk> new_brk >= current_pcb->addr_limit) if (new_brk < mm->brk_start || new_brk > new_brk >= current_pcb->addr_limit)
return mm->brk_end; return mm->brk_end;
if (mm->brk_end == new_brk) if (mm->brk_end == new_brk)
return new_brk; return new_brk;
@ -392,9 +407,6 @@ uint64_t sys_pipe(struct pt_regs *regs)
extern uint64_t sys_mkdir(struct pt_regs *regs); extern uint64_t sys_mkdir(struct pt_regs *regs);
extern int sys_dup(int oldfd);
extern int sys_dup2(int oldfd, int newfd);
system_call_t system_call_table[MAX_SYSTEM_CALL_NUM] = { system_call_t system_call_table[MAX_SYSTEM_CALL_NUM] = {
[0] = system_call_not_exists, [0] = system_call_not_exists,
[1] = sys_put_string, [1] = sys_put_string,
@ -426,5 +438,18 @@ system_call_t system_call_table[MAX_SYSTEM_CALL_NUM] = {
[27] = sys_sched, [27] = sys_sched,
[28] = sys_dup, [28] = sys_dup,
[29] = sys_dup2, [29] = sys_dup2,
[30 ... 255] = system_call_not_exists, [30] = sys_socket,
[31] = sys_setsockopt,
[32] = sys_getsockopt,
[33] = sys_connect,
[34] = sys_bind,
[35] = sys_sendto,
[36] = sys_recvfrom,
[37] = sys_recvmsg,
[38] = sys_listen,
[39] = sys_shutdown,
[40] = sys_accept,
[41] = sys_getsockname,
[42] = sys_getpeername,
[43 ... 255] = system_call_not_exists,
}; };

View File

@ -41,5 +41,19 @@
#define SYS_SCHED 27 // 让系统立即运行调度器该系统调用不能由运行在Ring3的程序发起 #define SYS_SCHED 27 // 让系统立即运行调度器该系统调用不能由运行在Ring3的程序发起
#define SYS_DUP 28 #define SYS_DUP 28
#define SYS_DUP2 29 #define SYS_DUP2 29
#define SYS_SOCKET 30 // 创建一个socket
#define SYS_SETSOCKOPT 31 // 设置socket的选项
#define SYS_GETSOCKOPT 32 // 获取socket的选项
#define SYS_CONNECT 33 // 连接到一个socket
#define SYS_BIND 34 // 绑定一个socket
#define SYS_SENDTO 35 // 向一个socket发送数据
#define SYS_RECVFROM 36 // 从一个socket接收数据
#define SYS_RECVMSG 37 // 从一个socket接收消息
#define SYS_LISTEN 38 // 监听一个socket
#define SYS_SHUTDOWN 39 // 关闭socket
#define SYS_ACCEPT 40 // 接受一个socket连接
#define SYS_GETSOCKNAME 41 // 获取socket的名字
#define SYS_GETPEERNAME 42 // 获取socket的对端名字
#define SYS_AHCI_END_REQ 255 // AHCI DMA请求结束end_request的系统调用 #define SYS_AHCI_END_REQ 255 // AHCI DMA请求结束end_request的系统调用

View File

@ -1,4 +1,4 @@
use core::sync::atomic::{AtomicBool, Ordering}; use core::sync::atomic::{compiler_fence, AtomicBool, Ordering};
use alloc::{ use alloc::{
boxed::Box, boxed::Box,
@ -75,15 +75,21 @@ impl Timer {
/// @brief 将定时器插入到定时器链表中 /// @brief 将定时器插入到定时器链表中
pub fn activate(&self) { pub fn activate(&self) {
let timer_list = &mut TIMER_LIST.lock();
let inner_guard = self.0.lock(); let inner_guard = self.0.lock();
let timer_list = &mut TIMER_LIST.lock();
// 链表为空,则直接插入 // 链表为空,则直接插入
if timer_list.is_empty() { if timer_list.is_empty() {
// FIXME push_timer // FIXME push_timer
timer_list.push_back(inner_guard.self_ref.upgrade().unwrap()); timer_list.push_back(inner_guard.self_ref.upgrade().unwrap());
drop(inner_guard);
drop(timer_list);
compiler_fence(Ordering::SeqCst);
return; return;
} }
let mut split_pos: usize = 0; let mut split_pos: usize = 0;
for (pos, elt) in timer_list.iter().enumerate() { for (pos, elt) in timer_list.iter().enumerate() {
if elt.0.lock().expire_jiffies > inner_guard.expire_jiffies { if elt.0.lock().expire_jiffies > inner_guard.expire_jiffies {
@ -94,6 +100,8 @@ impl Timer {
let mut temp_list: LinkedList<Arc<Timer>> = timer_list.split_off(split_pos); let mut temp_list: LinkedList<Arc<Timer>> = timer_list.split_off(split_pos);
timer_list.push_back(inner_guard.self_ref.upgrade().unwrap()); timer_list.push_back(inner_guard.self_ref.upgrade().unwrap());
timer_list.append(&mut temp_list); timer_list.append(&mut temp_list);
drop(inner_guard);
drop(timer_list);
} }
#[inline] #[inline]
@ -147,20 +155,38 @@ impl SoftirqVec for DoTimerSoftirq {
// 最多只处理TIMER_RUN_CYCLE_THRESHOLD个计时器 // 最多只处理TIMER_RUN_CYCLE_THRESHOLD个计时器
for _ in 0..TIMER_RUN_CYCLE_THRESHOLD { for _ in 0..TIMER_RUN_CYCLE_THRESHOLD {
// kdebug!("DoTimerSoftirq run"); // kdebug!("DoTimerSoftirq run");
let timer_list = TIMER_LIST.try_lock();
let timer_list = &mut TIMER_LIST.lock(); if timer_list.is_err() {
continue;
}
let mut timer_list = timer_list.unwrap();
if timer_list.is_empty() { if timer_list.is_empty() {
break; break;
} }
if timer_list.front().unwrap().0.lock().expire_jiffies let timer_list_front = timer_list.pop_front().unwrap();
<= unsafe { TIMER_JIFFIES as u64 } // kdebug!("to lock timer_list_front");
{ let mut timer_list_front_guard = None;
let timer = timer_list.pop_front().unwrap(); for _ in 0..10 {
drop(timer_list); let x = timer_list_front.0.try_lock();
timer.run(); if x.is_err() {
continue;
}
timer_list_front_guard = Some(x.unwrap());
} }
if timer_list_front_guard.is_none() {
continue;
}
let timer_list_front_guard = timer_list_front_guard.unwrap();
if timer_list_front_guard.expire_jiffies > unsafe { TIMER_JIFFIES as u64 } {
drop(timer_list_front_guard);
timer_list.push_front(timer_list_front);
break;
}
drop(timer_list_front_guard);
drop(timer_list);
timer_list_front.run();
} }
self.clear_run(); self.clear_run();
@ -293,7 +319,7 @@ pub extern "C" fn rs_timer_next_n_us_jiffies(expire_us: u64) -> u64 {
pub extern "C" fn rs_timer_get_first_expire() -> i64 { pub extern "C" fn rs_timer_get_first_expire() -> i64 {
match timer_get_first_expire() { match timer_get_first_expire() {
Ok(v) => return v as i64, Ok(v) => return v as i64,
Err(e) => return e.to_posix_errno() as i64, Err(_) => return 0,
} }
} }

View File

@ -43,7 +43,7 @@ install_ubuntu_debian_pkg()
gnupg \ gnupg \
lsb-release \ lsb-release \
llvm-dev libclang-dev clang gcc-multilib \ llvm-dev libclang-dev clang gcc-multilib \
gcc build-essential fdisk dosfstools gcc build-essential fdisk dosfstools dnsmasq bridge-utils iptables
if [ -z "$(which docker)" ] && [ -n ${dockerInstall} ]; then if [ -z "$(which docker)" ] && [ -n ${dockerInstall} ]; then
echo "正在安装docker..." echo "正在安装docker..."

24
tools/qemu/ifdown-nat Executable file
View File

@ -0,0 +1,24 @@
#!/bin/bash
BRIDGE=dragonos-bridge
if [ -n "$1" ]; then
echo "正在断开接口 $1"
ip link set $1 down
brctl delif "$BRIDGE" $1
tap=`brctl show | grep natnet | awk '{print $4}'`
if [[ $tap != tap* ]];then
ip link set "$BRIDGE" down
brctl delbr "$BRIDGE"
iptables -t nat -F
kill `ps aux | grep dnsmasq | grep -v grep | awk '{print $2}'`
echo "断开接口 $1 成功"
echo "网桥 $BRIDGE 卸载成功"
echo "dnsmasq 服务停止成功"
exit 0
else
echo "断开接口 $1 成功"
exit 0
fi
else
echo "删除错误:未指定接口"
exit 1
fi

85
tools/qemu/ifup-nat Executable file
View File

@ -0,0 +1,85 @@
#!/bin/bash
# 设置 bridge 名称
BRIDGE=dragonos-bridge
# 设置网络信息
NETWORK=192.168.137.0
NETMASK=255.255.255.0
GATEWAY=192.168.137.1
DHCPRANGE=192.168.137.100,192.168.137.200
# 启用PXE支持的可选参数
TFTPROOT=
BOOTP=
function check_bridge()
{
if brctl show | grep "^$BRIDGE" &> /dev/null; then
return 1
else
return 0
fi
}
function create_bridge()
{
brctl addbr "$BRIDGE"
brctl stp "$BRIDGE" on
brctl setfd "$BRIDGE" 0
ifconfig "$BRIDGE" "$GATEWAY" netmask "$NETMASK" up
}
function enable_ip_forward()
{
echo 1 > /proc/sys/net/ipv4/ip_forward
}
function add_filter_rules()
{
iptables -t nat -A POSTROUTING -s "$NETWORK"/"$NETMASK" \
! -d "$NETWORK"/"$NETMASK" -j MASQUERADE
}
function start_dnsmasq()
{
# 禁止重复运行dnsmasq
ps -ef | grep "dnsmasq" | grep -v "grep" &> /dev/null
if [ $? -eq 0 ]; then
echo "dnsmasq 已经在运行"
return 1
fi
dnsmasq \
--strict-order \
--except-interface=lo \
--interface=$BRIDGE \
--listen-address=$GATEWAY \
--bind-interfaces \
--dhcp-range=$DHCPRANGE \
--conf-file="" \
--pid-file=/var/run/qemu-dhcp-$BRIDGE.pid \
--dhcp-leasefile=/var/run/qemu-dhcp-$BRIDGE.leases \
--dhcp-no-override \
${TFTPROOT:+"--enable-tftp"} \
${TFTPROOT:+"--tftp-root=$TFTPROOT"} \
${BOOTP:+"--dhcp-boot=$BOOTP"}
}
function setup_bridge_nat()
{
check_bridge "$BRIDGE"
if [ $? -eq 0 ]; then
create_bridge
fi
enable_ip_forward
add_filter_rules "$BRIDGE"
start_dnsmasq "$BRIDGE"
}
# 安装前需要检查$1参数
if [ -n "$1" ]; then
setup_bridge_nat
brctl addif "$BRIDGE" "$1"
ifconfig "$1" 0.0.0.0 up
exit 0
else
echo "发现错误:没有指定接口"
exit 1
fi

View File

@ -1,9 +1,40 @@
check_dependencies()
{
# Check if qemu is installed
if [ -z "$(which qemu-system-x86_64)" ]; then
echo "Please install qemu first!"
exit 1
fi
# Check if brctl is installed
if [ -z "$(which brctl)" ]; then
echo "Please install bridge-utils first!"
exit 1
fi
# Check if dnsmasq is installed
if [ -z "$(which dnsmasq)" ]; then
echo "Please install dnsmasq first!"
exit 1
fi
# Check if iptable is installed
if [ -z "$(which iptables)" ]; then
echo "Please install iptables first!"
exit 1
fi
}
check_dependencies
# 进行启动前检查 # 进行启动前检查
flag_can_run=1 flag_can_run=1
ARGS=`getopt -o p -l bios:,display: -- "$@"` ARGS=`getopt -o p -l bios:,display: -- "$@"`
eval set -- "${ARGS}" eval set -- "${ARGS}"
echo "$@" echo "$@"
allflags=$(qemu-system-x86_64 -cpu help | awk '/flags/ {y=1; getline}; y {print}' | tr ' ' '\n' | grep -Ev "^$" | sed -r 's|^|+|' | tr '\n' ',' | sed -r "s|,$||") allflags=
# allflags=$(qemu-system-x86_64 -cpu help | awk '/flags/ {y=1; getline}; y {print}' | tr ' ' '\n' | grep -Ev "^$" | sed -r 's|^|+|' | tr '\n' ',' | sed -r "s|,$||")
ARCH="x86_64" ARCH="x86_64"
#ARCH="i386" #ARCH="i386"
# 请根据自己的需要,在-d 后方加入所需的 trace 事件 # 请根据自己的需要,在-d 后方加入所需的 trace 事件
@ -17,7 +48,6 @@ if [ $(uname) == Darwin ]; then
qemu_accel=hvf qemu_accel=hvf
fi fi
QEMU=qemu-system-x86_64 QEMU=qemu-system-x86_64
QEMU_DISK_IMAGE="../bin/disk.img" QEMU_DISK_IMAGE="../bin/disk.img"
QEMU_MEMORY="512M" QEMU_MEMORY="512M"
@ -29,7 +59,10 @@ QEMU_RTC_CLOCK="clock=host,base=localtime"
QEMU_SERIAL="file:../serial_opt.txt" QEMU_SERIAL="file:../serial_opt.txt"
QEMU_DRIVE="id=disk,file=${QEMU_DISK_IMAGE},if=none" QEMU_DRIVE="id=disk,file=${QEMU_DISK_IMAGE},if=none"
QEMU_DEVICES="-device ahci,id=ahci -device ide-hd,drive=disk,bus=ahci.0 -nic user,model=virtio-net-pci -usb -device qemu-xhci,id=xhci,p2=8,p3=4 -machine accel=${qemu_accel} -machine q35"
# ps: 下面这条使用tap的方式无法dhcp获取到ip暂时不知道为什么
# QEMU_DEVICES="-device ahci,id=ahci -device ide-hd,drive=disk,bus=ahci.0 -net nic,netdev=nic0 -netdev tap,id=nic0,model=virtio-net-pci,script=qemu/ifup-nat,downscript=qemu/ifdown-nat -usb -device qemu-xhci,id=xhci,p2=8,p3=4 -machine accel=${qemu_accel} -machine q35 "
QEMU_DEVICES="-device ahci,id=ahci -device ide-hd,drive=disk,bus=ahci.0 -nic user,model=virtio-net-pci,hostfwd=tcp::12580-:12580 -usb -device qemu-xhci,id=xhci,p2=8,p3=4 -machine accel=${qemu_accel} -machine q35 "
QEMU_ARGUMENT="-d ${QEMU_DISK_IMAGE} -m ${QEMU_MEMORY} -smp ${QEMU_SMP} -boot order=d -monitor ${QEMU_MONITOR} -d ${qemu_trace_std} " QEMU_ARGUMENT="-d ${QEMU_DISK_IMAGE} -m ${QEMU_MEMORY} -smp ${QEMU_SMP} -boot order=d -monitor ${QEMU_MONITOR} -d ${qemu_trace_std} "

View File

@ -1,34 +1,242 @@
/** #include <arpa/inet.h>
* @file main.c #include <fcntl.h>
* @author longjin (longjin@RinGoTek.cn)
* @brief signal用的程序
* @version 0.1
* @date 2022-12-06
*
* @copyright Copyright (c) 2022
*
*/
/**
* signal的kill命令的方法:
* 1.DragonOS的控制台输入 exec bin/test_signal.elf &
* , '&',
* 2.kill对应的进程的pid (,:"[1] 生成的pid")
*
*/
#include <math.h>
#include <signal.h> #include <signal.h>
#include <stdio.h> #include <stdio.h>
#include <stdlib.h> #include <stdlib.h>
#include <time.h> #include <string.h>
#include <unistd.h> #include <unistd.h>
#define CONN_QUEUE_SIZE 20
#define BUFFER_SIZE 1024
#define SERVER_PORT 12580
int main() int server_sockfd;
int conn;
void signal_handler(int signo)
{ {
printf("Test Relibc printf!\n");
printf("Test Relibc printf ok!\n"); printf("Server is exiting...\n");
close(conn);
close(server_sockfd);
exit(0);
}
static char logo[] =
" ____ ___ ____ \n| _ \\ _ __ __ _ __ _ ___ _ __ / _ \\ / ___| "
"\n| | | || '__| / _` | / _` | / _ \\ | '_ \\ | | | |\\___ \\ \n| |_| || | | (_| || (_| || (_) || | | || |_| | "
"___) |\n|____/ |_| \\__,_| \\__, | \\___/ |_| |_| \\___/ |____/ \n |___/ \n";
void tcp_server()
{
printf("TCP Server is running...\n");
server_sockfd = socket(AF_INET, SOCK_STREAM, 0);
printf("socket() ok, server_sockfd=%d\n", server_sockfd);
struct sockaddr_in server_sockaddr;
server_sockaddr.sin_family = AF_INET;
server_sockaddr.sin_port = htons(SERVER_PORT);
server_sockaddr.sin_addr.s_addr = htonl(INADDR_ANY);
if (bind(server_sockfd, (struct sockaddr *)&server_sockaddr, sizeof(server_sockaddr)))
{
perror("Server bind error.\n");
exit(1);
}
printf("TCP Server is listening...\n");
if (listen(server_sockfd, CONN_QUEUE_SIZE) == -1)
{
perror("Server listen error.\n");
exit(1);
}
printf("listen() ok\n");
char buffer[BUFFER_SIZE];
struct sockaddr_in client_addr;
socklen_t client_length = sizeof(client_addr);
/*
Await a connection on socket FD.
When a connection arrives, open a new socket to communicate with it,
set *ADDR (which is *ADDR_LEN bytes long) to the address of the connecting
peer and *ADDR_LEN to the address's actual length, and return the
new socket's descriptor, or -1 for errors.
*/
conn = accept(server_sockfd, (struct sockaddr *)&client_addr, &client_length);
printf("Connection established.\n");
if (conn < 0)
{
printf("Create connection failed, code=%d\n", conn);
exit(1);
}
send(conn, logo, sizeof(logo), 0);
while (1)
{
memset(buffer, 0, sizeof(buffer));
int len = recv(conn, buffer, sizeof(buffer), 0);
if (len <= 0)
{
printf("Receive data failed! len=%d\n", len);
break;
}
if (strcmp(buffer, "exit\n") == 0)
{
break;
}
printf("Received: %s\n", buffer);
send(conn, buffer, len, 0);
}
close(conn);
close(server_sockfd);
}
void udp_server()
{
printf("UDP Server is running...\n");
server_sockfd = socket(AF_INET, SOCK_DGRAM, 0);
printf("socket() ok, server_sockfd=%d\n", server_sockfd);
struct sockaddr_in server_sockaddr;
server_sockaddr.sin_family = AF_INET;
server_sockaddr.sin_port = htons(SERVER_PORT);
server_sockaddr.sin_addr.s_addr = htonl(INADDR_ANY);
if (bind(server_sockfd, (struct sockaddr *)&server_sockaddr, sizeof(server_sockaddr)))
{
perror("Server bind error.\n");
exit(1);
}
printf("UDP Server is listening...\n");
char buffer[BUFFER_SIZE];
struct sockaddr_in client_addr;
socklen_t client_length = sizeof(client_addr);
while (1)
{
memset(buffer, 0, sizeof(buffer));
int len = recvfrom(server_sockfd, buffer, sizeof(buffer), 0, (struct sockaddr *)&client_addr, &client_length);
if (len <= 0)
{
printf("Receive data failed! len=%d", len);
break;
}
if (strcmp(buffer, "exit\n") == 0)
{
break;
}
printf("Received: %s", buffer);
sendto(server_sockfd, buffer, len, 0, (struct sockaddr *)&client_addr, client_length);
printf("Send: %s", buffer);
}
close(conn);
close(server_sockfd);
}
void tcp_client()
{
printf("Client is running...\n");
int client_sockfd = socket(AF_INET, SOCK_STREAM, 0);
struct sockaddr_in server_addr = {0};
server_addr.sin_family = AF_INET;
server_addr.sin_port = htons(12581);
server_addr.sin_addr.s_addr = inet_addr("192.168.199.129");
printf("to connect\n");
if (connect(client_sockfd, (struct sockaddr *)&server_addr, sizeof(server_addr)) < 0)
{
perror("Failed to establish connection to server\n");
exit(1);
}
printf("connected to server\n");
char sendbuf[BUFFER_SIZE] = {0};
char recvbuf[BUFFER_SIZE] = {0};
int x = recv(client_sockfd, recvbuf, sizeof(recvbuf), 0);
fputs(recvbuf, stdout);
memset(recvbuf, 0, sizeof(recvbuf));
while (1)
{
fgets(sendbuf, sizeof(sendbuf), stdin);
sendbuf[0] = 'a';
// printf("to send\n");
send(client_sockfd, sendbuf, strlen(sendbuf), 0);
// printf("send ok\n");
if (strcmp(sendbuf, "exit\n") == 0)
{
break;
}
int x = recv(client_sockfd, recvbuf, sizeof(recvbuf), 0);
if (x < 0)
{
printf("recv error, retval=%d\n", x);
break;
}
fputs(recvbuf, stdout);
memset(recvbuf, 0, sizeof(recvbuf));
memset(sendbuf, 0, sizeof(sendbuf));
}
close(client_sockfd);
}
void udp_client()
{
struct sockaddr_in addr;
int sockfd, len = 0;
int addr_len = sizeof(struct sockaddr_in);
char buffer[256];
/* 建立socket注意必须是SOCK_DGRAM */
if ((sockfd = socket(AF_INET, SOCK_DGRAM, 0)) < 0)
{
perror("socket");
exit(1);
}
/* 填写sockaddr_in*/
bzero(&addr, sizeof(addr));
addr.sin_family = AF_INET;
addr.sin_port = htons(12581);
addr.sin_addr.s_addr = inet_addr("192.168.199.129");
printf("to send logo\n");
sendto(sockfd, logo, sizeof(logo), 0, (struct sockaddr *)&addr, addr_len);
printf("send logo ok\n");
while (1)
{
bzero(buffer, sizeof(buffer));
printf("Please enter a string to send to server: \n");
/* 从标准输入设备取得字符串*/
len = read(STDIN_FILENO, buffer, sizeof(buffer));
printf("to send: %d\n", len);
/* 将字符串传送给server端*/
sendto(sockfd, buffer, len, 0, (struct sockaddr *)&addr, addr_len);
/* 接收server端返回的字符串*/
len = recvfrom(sockfd, buffer, sizeof(buffer), 0, (struct sockaddr *)&addr, &addr_len);
printf("Receive from server: %s\n", buffer);
}
return 0; return 0;
}
void main()
{
// signal(SIGKILL, signal_handler);
// signal(SIGINT, signal_handler);
tcp_server();
// udp_server();
// tcp_client();
// udp_client();
} }

View File

@ -34,6 +34,20 @@
#define SYS_GETPID 26 // 获取当前进程的pid进程标识符 #define SYS_GETPID 26 // 获取当前进程的pid进程标识符
#define SYS_DUP 28 #define SYS_DUP 28
#define SYS_DUP2 29 #define SYS_DUP2 29
#define SYS_SOCKET 30 // 创建一个socket
#define SYS_SETSOCKOPT 31 // 设置socket的选项
#define SYS_GETSOCKOPT 32 // 获取socket的选项
#define SYS_CONNECT 33 // 连接到一个socket
#define SYS_BIND 34 // 绑定一个socket
#define SYS_SENDTO 35 // 向一个socket发送数据
#define SYS_RECVFROM 36 // 从一个socket接收数据
#define SYS_RECVMSG 37 // 从一个socket接收消息
#define SYS_LISTEN 38 // 监听一个socket
#define SYS_SHUTDOWN 39 // 关闭socket
#define SYS_ACCEPT 40 // 接受一个socket连接
#define SYS_GETSOCKNAME 41 // 获取socket的名字
#define SYS_GETPEERNAME 42 // 获取socket的对端名字
/** /**
* @brief * @brief