/freebsd/sys/contrib/openzfs/.github/workflows/scripts/ |
H A D | qemu-3-deps.sh | 2 # 3) Wait for VM to boot from previous step and launch dependencies 10 .github/workflows/scripts/qemu-wait-for-vm.sh vm0 18 if [ -n "${2:-}" ] ; then 19 sed -i -E 's/Linux-Maximum: .+/Linux-Maximum: 99.99/g' META 22 scp .github/workflows/scripts/qemu-3-deps-vm.sh zfs@vm0:qemu-3-deps-vm.sh 23 PID=`pidof /usr/bin/qemu-system-x86_64` 24 ssh zfs@vm0 '$HOME/qemu-3-deps-vm.sh' "$@" 26 tail --pid=$PID -f /dev/null 28 rm -f $HOME/.ssh/known_hosts
|
H A D | qemu-6-tests.sh | 6 # called on runner: qemu-6-tests.sh 7 # called on qemu-vm: qemu-6-tests.sh $OS $2/$3 10 set -eu 17 DIFF=$((CURRENT-TSSTART)) 19 DIFF=$((DIFF-(H*3600))) 21 S=$((DIFF-(M*60))) 24 echo $LINE| grep -q "^Test[: ]" && CTR=$((CTR+1)) && echo $CTR > /tmp/ctr 27 COLOR="$BASE/scripts/zfs-tests-color.sh" 28 CLINE=$(echo $LINE| grep "^Test[ :]" | sed -e 's|/usr/local|/usr|g' \ 29 | sed -e 's| /usr/share/zfs/zfs-tests/tests/| |g' | $COLOR) [all …]
|
H A D | qemu-4-build.sh | 6 echo "Build modules in QEMU machine" 8 # Bring our VM back up and copy over ZFS source 9 .github/workflows/scripts/qemu-prepare-for-build.sh 11 ssh zfs@vm0 '$HOME/zfs/.github/workflows/scripts/qemu-4-build-vm.sh' $@
|
H A D | qemu-2-start.sh | 4 # 2) start qemu with some operating system, init via cloud-init 7 set -eu 9 # short name used in zfs-qemu.yml 12 # OS variant (virt-install --os-variant list) 16 FREEBSD_REL="https://download.freebsd.org/releases/CI-IMAGES" 17 FREEBSD_SNAP="https://download.freebsd.org/snapshots/CI-IMAGES" 21 UBMIRROR="https://cloud-images.ubuntu.com" 22 #UBMIRROR="https://mirrors.cloud.tencent.com/ubuntu-cloud-images" 23 #UBMIRROR="https://mirror.citrahost.com/ubuntu-cloud-images" 25 # default nic model for vm's [all …]
|
H A D | qemu-prepare-for-build.sh | 3 # Helper script to run after installing dependencies. This brings the VM back 5 echo "Build modules in QEMU machine" 7 .github/workflows/scripts/qemu-wait-for-vm.sh vm0 8 rsync -ar $HOME/work/zfs/zfs zfs@vm0:./
|
H A D | qemu-5-setup.sh | 7 set -eu 13 PID=$(pidof /usr/bin/qemu-system-x86_64) 14 tail --pid=$PID -f /dev/null 34 # setup the testing vm's 39 echo "Creating disk for vm$i..." 40 DISK="/dev/zvol/zpool/vm$i" 42 sudo zfs clone zpool/openzfs@now zpool/vm$i-system 43 sudo zfs create -ps -b 64k -V 64g zpool/vm$i-tests 45 cat <<EOF > /tmp/user-data 46 #cloud-config [all …]
|
H A D | qemu-7-prepare.sh | 5 # - this script pre-creates all needed logfiles for later summary 8 set -eu 14 mkdir -p $RESPATH 17 if [ -z ${VMs:-} ]; then 22 tar cf /tmp/qemu-$OS.tar -C $RESPATH -h . || true 30 # catch result files of testings (vm's should be there) 32 rsync -arL zfs@vm$i:$RESPATH/current $RESPATH/vm$i || true 33 scp zfs@vm$i:"/var/tmp/*.txt" $RESPATH/vm$i || true 34 scp zfs@vm$i:"/var/tmp/*.rpm" $RESPATH/vm$i || true 36 cp -f /var/tmp/*.txt $RESPATH || true [all …]
|
H A D | qemu-wait-for-vm.sh | 3 # Wait for a VM to boot up and become active. This is used in a number of our 6 # $1: VM hostname or IP address 8 while pidof /usr/bin/qemu-system-x86_64 >/dev/null; do 9 ssh 2>/dev/null zfs@$1 "uname -a" && break
|
H A D | qemu-9-summary-page.sh | 7 set -eu 10 echo -e $* >> "out-$logfile.md" 14 cat "$1" >> "out-$logfile.md" 19 cat "$1" >> "out-$logfile.md" 24 test -f "$1" || exit 0 28 …/docs.github.com/en/enterprise-server@3.6/actions/using-workflows/workflow-commands-for-github-act… 34 if [ ! -f out-1.md ]; then 36 for tarfile in Logs-functional-*/qemu-*.tar; do 37 rm -rf vm* *.txt 38 if [ ! -s "$tarfile" ]; then [all …]
|
H A D | qemu-1-setup.sh | 4 # 1) setup qemu instance on action runner 7 set -eu 11 sudo apt-get -y update 12 sudo apt-get install -y axel cloud-image-utils daemonize guestfs-tools \ 13 virt-manager linux-modules-extra-$(uname -r) zfsutils-linux 16 rm -f ~/.ssh/id_ed25519 17 ssh-keygen -t ed25519 -f ~/.ssh/id_ed25519 -q -N "" 24 sudo swapoff -a 25 sudo umount -l /mnt 26 DISK="/dev/disk/cloud/azure_resource-part1" [all …]
|
H A D | qemu-test-repo-vm.sh | 7 # ./qemu-test-repo-vm [URL] 10 # If blank, use the default repo from zfs-release RPM. 12 set -e 14 source /etc/os-release 19 if [ -n "$1" ] ; then 25 SUMMARY=/tmp/repo/$OS-$VERSION-summary.txt 27 # $1: Repo 'zfs' 'zfs-kmod' 'zfs-testing' 'zfs-testing-kmod' 29 # install from. Blank means use default from zfs-release RPM. 33 if [ -n "$2" ] ; then 37 args="--disablerepo=zfs --enablerepo=$repo" [all …]
|
/freebsd/tools/boot/ |
H A D | ci-qemu-test.sh | 4 # in QEMU and echo "Hello world." from init, as a very quick smoke test for CI. 5 # Uses QEMU's virtual FAT filesystem to avoid the need to create a disk image. 7 # as a quick smoke-test as long as pkgbase packages have been built. The 12 set -e 22 trap - EXIT SIGINT SIGHUP SIGTERM SIGQUIT 23 rm -rf ${WORKDIR} 31 mkdir -p ${ROOTDIR}/${dir} 40 -o ABI_FILE=$OBJTOP/bin/sh/sh \ 41 -C ${ROOTDIR}/pkg.conf -r ${ROOTDIR} install \ 42 FreeBSD-kernel-generic FreeBSD-bootloader \ [all …]
|
/freebsd/sys/contrib/openzfs/.github/workflows/ |
H A D | zfs-qemu.yml | 1 name: zfs-qemu 22 …tion: "(optional) Experimental kernel version to install on Fedora (like '6.14' or '6.13.3-0.rc3')" 25 group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }} 26 cancel-in-progress: true 29 test-config: 31 runs-on: ubuntu-24.04 36 - uses: actions/checkout@v4 38 fetch-depth: 0 39 - name: Generate OS config and CI type 42 …, "debian11", "debian12", "fedora41", "fedora42", "freebsd13-4r", "freebsd14-3s", "freebsd15-0c", … [all …]
|
H A D | zfs-qemu-packages.yml | 6 # Build RPMs - Build release RPMs and tarballs and put them into an artifact 10 # Test repo - Test install the ZFS RPMs from the ZFS repo. On EL distos, this 17 # zfs-release RPM (http://download.zfsonlinux.org). 22 name: zfs-qemu-packages 33 - "Build RPMs" 34 - "Test repo" 41 group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }} 42 cancel-in-progress: true 45 zfs-qemu-packages-jobs: 46 name: qemu-VMs [all …]
|
H A D | checkstyle.yaml | 8 group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }} 9 cancel-in-progress: true 13 runs-on: ubuntu-22.04 15 - uses: actions/checkout@v4 18 - name: Install dependencies 21 sudo apt-get purge -y snapd google-chrome-stable firefox 22 ONLY_DEPS=1 .github/workflows/scripts/qemu-3-deps-vm.sh ubuntu22 23 sudo apt-get install -y cppcheck devscripts mandoc pax-utils shellcheck 24 sudo python -m pipx install --quiet flake8 27 checkbashisms --version [all …]
|
H A D | zloop.yml | 8 group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }} 9 cancel-in-progress: true 13 runs-on: ubuntu-24.04 17 - uses: actions/checkout@v4 20 - name: Install dependencies 22 sudo apt-get purge -y snapd google-chrome-stable firefox 23 ONLY_DEPS=1 .github/workflows/scripts/qemu-3-deps-vm.sh ubuntu24 24 - name: Autogen.sh 26 sed -i '/DEBUG_CFLAGS="-Werror"/s/^/#/' config/zfs-build.m4 28 - name: Configure [all …]
|
/freebsd/release/ |
H A D | Makefile.vm | 5 VMTARGETS= vm-image 15 VMBASE?= vm 17 VHD_DESC= Azure, VirtualPC, Hyper-V, Xen disk image 19 QCOW2_DESC= Qemu, KVM disk image 24 BASIC-CI 27 BASIC-CI \ 28 BASIC-CLOUDINIT \ 37 BASIC-CI_FORMAT= raw 38 BASIC-CI_FSLIST?= ufs 39 BASIC-CI_DESC= Image for CI [all …]
|
/freebsd/tests/ci/ |
H A D | Makefile | 1 # SPDX-License-Identifier: BSD-2-Clause 10 # User-driven targets: 12 # ci-smoke: Run smoke tests which is simply booting the image 13 # ci-full: Run full tests 18 # USE_QEMU: Use QEMU for testing rather than bhyve 25 _MEMORY!=sysctl -n hw.physmem 2>/dev/null 26 PARALLEL_JOBS!=sysctl -n hw.ncpu 2>/dev/null || nproc 2>/dev/null 57 OSRELEASE= ${TYPE}-${REVISION}-${BRANCH}-${TARGET} 58 VOLUME_LABEL= ${REVISION:C/[.-]/_/g}_${BRANCH:C/[.-]/_/g}_${TARGET} 60 OSRELEASE= ${TYPE}-${REVISION}-${BRANCH}-${TARGET}-${TARGET_ARCH} [all …]
|
/freebsd/usr.sbin/bhyve/ |
H A D | bhyve.8 | 99 I/O connectivity can be specified with command-line parameters. 110 .Pa edk2-bhyve 113 .Pa u-boot-bhyve-arm64 114 package provides a U-Boot image that can be used to boot the guest. 120 .Bl -tag -width 10n 161 Destroy the VM on guest initiated power-off. 203 Set configuration variables from a simple, key-value config file. 221 When using a VNC client that supports QEMU Extended Key Event Message (e.g. 223 When using a VNC client that doesn't support QEMU Extended Key Event Message 229 Allow devices behind the LPC PCI-ISA bridge to be configured. [all …]
|
H A D | bhyve_config.5 | 1 .\" SPDX-License-Identifier: BSD-2-Clause 35 per-device settings. 69 For those variables the following case-insensitive values may be used to 72 .Bl -bullet -offset indent -compact 85 .Bl -bullet -offset indent -compact 103 .Bl -column "memory.guest_in_core" "integer" "Default" 106 The name of the VM. 140 This can cause problems if the guest uses the in-memory version, since certain 144 Destroy the VM on guest-initiated power-off. 149 If this is set to a non-zero value, a debug server [all …]
|
/freebsd/sys/arm/qemu/ |
H A D | virt_mp.c | 1 /*- 34 #include <vm/vm.h> 35 #include <vm/pmap.h> 45 #include <arm/qemu/virt_mp.h>
|
H A D | virt_machdep.c | 1 /*- 34 #include <vm/vm.h> 40 #include <arm/qemu/virt_mp.h> 65 FDT_PLATFORM_DEF(virt, "virt", 0, "linux,dummy-virt", 1);
|
/freebsd/share/man/man4/ |
H A D | ptnet.4 | 30 .Nd Ethernet driver for passed-through netmap ports 36 .Bd -ragged -offset indent 43 from within a Virtual Machine (VM). 45 the VM can access the TX/RX rings and buffers of a netmap port 46 that the hypervisor has passed-through to the VM. 49 is currently available for QEMU/KVM. 52 port can be passed-through, including physical NICs, 56 The main use-case for netmap passthrough is Network Function 58 VMs may want to process very high packet rates (e.g., 1-10 millions 68 comes from the hypervisor being completely bypassed in the data-path. [all …]
|
/freebsd/sys/powerpc/booke/ |
H A D | machdep_e500.c | 1 /*- 2 * SPDX-License-Identifier: BSD-2-Clause 4 * Copyright (c) 2011-2012 Semihalf. 36 #include <vm/vm.h> 37 #include <vm/pmap.h> 58 /* Enable D-cache if applicable */ in booke_enable_l1_cache() 67 printf("L1 D-cache %sabled\n", in booke_enable_l1_cache() 70 /* Enable L1 I-cache if applicable. */ in booke_enable_l1_cache() 79 printf("L1 I-cache %sabled\n", in booke_enable_l1_cache() 99 * This fixes waiting forever for cache enable in qemu, in booke_enable_l2_cache()
|
/freebsd/sys/dev/vmgenc/ |
H A D | vmgenc_acpi.c | 1 /*- 2 * SPDX-License-Identifier: BSD-2-Clause 29 * VM Generation Counter driver 33 * https://docs.microsoft.com/en-us/windows/win32/hyperv_v2/virtual-machine-generation-identifier , 34 * https://azure.microsoft.com/en-us/blog/accessing-and-using-azure-vm-unique-id/ 39 * - QEMU: https://bugzilla.redhat.com/show_bug.cgi?id=1118834 40 * - VMware/ESXi: https://kb.vmware.com/s/article/2032586 41 …* - Xen: https://github.com/xenserver/xen-4.5/blob/master/tools/firmware/hvmloader/acpi/dsdt.asl#L… 73 MODULE_PNP_INFO("Z:_CID", acpi, vmgenc, vmgenc_ids, nitems(vmgenc_ids) - 1); 88 sizeof(((struct harvest_event *)0)->he_entropy)); in vmgenc_harvest_all() [all …]
|