Giter Site home page Giter Site logo

visionos_30days's People

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

visionos_30days's Issues

I have a question

Hello, I am a student interested in visionOS development. I would like to ask you for advice on how to use the eye tracking function in visionOS development. At present, I don’t seem to have found the relevant interface calling instructions.

Whats the plan?

Is this for personal use? I have enjoyed following along and participating.

what is worldtracking meaning

I run your demo day04
Task {
try await session.run([worldTracking])

        for await update in worldTracking.anchorUpdates {
            switch update.event {
            case .added, .updated:
                print("Anchor position updated.")
            case .removed:
                print("Anchor position now unknown.")
            @unknown default:
                break
            }
        }
    }

Is it possible to collect eye movement data?

Hi,

Thank you for your awesome work.
I know Vision Pro has eye-tracking. I just wondering whether and how the developer can acquire and collect eye-tracking data using this device?

Best.

Day 22

Hi,

I was looking at the same issue yesterday, and found you had a solution as well.

Are you using the paid google street view api? Would you explain a bit more of the code? Thanks!
And is there a way to get the panoid according to the latitude and longitude?

var baseUrl = URL(string: "https://streetviewpixels-pa.googleapis.com/v1/tile")!
let paramsCommon: [URLQueryItem] = [
URLQueryItem(name: "cb_client", value: "maps_sv.tactile"),
URLQueryItem(name: "panoid", value: panoId),
URLQueryItem(name: "zoom", value: "(3)")
]
baseUrl.append(queryItems: paramsCommon)

World Sensing

Hello,

I run your Day4. in XCod 15.2 Beta and on a simulator.

  1. It seems authorization request doesn't pop up on my environment.
  2. In your Day4 Video, it seems it recognizes wall / floor etc.. but
    I don't see that happening. Perhaps, authorization won't pop up?

I see below printout.

WorldTrackingProvider.isSupported: true
PlaneDetectionProvider.isSupported: false
SceneReconstructionProvider.isSupported: false
HandTrackingProvider.isSupported: false

If you have any idea how to enable functionality, please let me know

Hand tracking for custom gestures

Hello, thanks for the great examples - this has been helpful.

Do you have plans to add / explore the new capabilities in ARKit around hand tracking? I'm curious if it's possible to run the simulator for hand-tracking without having the apple vision pro device itself - what is that experience like? I'm having trouble finding additional information about it online. Thanks!

Day 4 example shows nothing

I've read about World Tracking not being in the simulator yet, but the screenshots in the README.md show it to be working.
Is this a screenshot from Apple or is it somehow working for you?

Day 5 - tap gesture not triggered

Hi,
Awesome project! Helped me a lot getting into and familiar with VisionPro and the SDK. However, one thing I struggled with was that all my SpatialTapGesture() didn't work. So, I looking into Day 5 and it wasn't working there either.. I eventually found Day 13 and this WWDC video which cleared things up a lot..
TL;DR, you need to set InputTargetComponent & CollisionComponent, which isn't being done for contentEntity.

So, I'd suggest either removing the .gesture or setting the components. Otherwise it's a bit confusing, at least imo.. ^^

Day 13 Missing axis.usdz file

Day 13 shows how to add an x/y/z axis to the immersive space. To do so, it attempts to load an axis.usdz file from the project's assets. This file is not included in the repo.

Where can we get this?

non-technical question

Brother, I would like to ask you a non-technical question, what is your motivation to learn the development of vision os, are you optimistic about the market of vision pro, or completely out of interest.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.