Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Documentation Improvements #4

Open
nebulorum opened this issue May 3, 2022 · 3 comments
Open

Documentation Improvements #4

nebulorum opened this issue May 3, 2022 · 3 comments

Comments

@nebulorum
Copy link

Congratulations on a great idea. I was looking for experiences on using LCD for etching and solder masking. This is a validation that it makes sense. Just going through the project I found some minor documentations issues:

  • The BOM has link for the LCD no longer works, it would be nice to have the spec of the item.
  • It's not clear how to build the project or run it as a developer.

I would also like to test the software before building something like this. I also wonder if I couldn't run the software on my PC and drive the display from it.

Anyway great work.

@nebulorum
Copy link
Author

nebulorum commented May 4, 2022

Was really curious and decided to clone and try to run. I managed with the following:

  1. to get is started without config-dev.json and the CNC disabled.
  2. change the screen resolution
  3. Load a file, for some reason until I had the UI correctly configured I was failed to see the files in the list. Once I ran the UI right everything clicked.

With the Arduino sample board from tracespace the images where too small, with a board of my own I could see it well. However I then can't go beyond because I don't know how to mock the exposure.

@joelkoz
Copy link
Owner

joelkoz commented May 8, 2022

Thanks for checking out the project. Build instructions are actually on the hackaday page here: https://hackaday.io/project/182645-qwickfab-pcb

Re: running from a PC instead of a pi - yes, you can. I've actually done that during development. The system when first checked out of github assumes it will be set up on a Raspberry Pi with a Sharp LS055R1SX04 LCD with HDMI converted as display 1, and a 480 x 320 SPI LCD touchscreen configured as display two. The Pi is configured in a multi-head mode which makes the display space look like on big canvas. Render images on the left side, and it appears on the Sharp LCD. Draw on the right side, and it appears on the SPI LCD.

The app is actually an Electron app, so the app window is just one large Chrome window that covers this entire screen area. This is important to know, because the UI and trace windows are layed out using HTML found in app/render/index.html and CSS found in app/render/ui/ui.css There are two primary <div> areas: one with an id of mask-area that contains a <canvas> with an id of uv-mask for the traces, and one with an id of ui. The CSS positions these areas in such a way that uv-mask ends up displaying on the Sharp LCD, and the ui area is on the small display. Assuming you know HTML and CSS, you can modify the CSS of ui.css to re-position the displays so the UI would appear on your PC screen, and the uv-mask would render on the LCD screen of the device you build.

@nebulorum
Copy link
Author

I did some digit into the Pi setup and kind of inferred there was a giant canvas. It also seems you can run separate screens. Not sure how to do the same on Mac, but I can investigate.

Is there a way to sale down the Display for testing (so I get the entire screen on my viewport)? I find using the virtual buttons hard since most of the screen is out of my view port.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants