الخميس، 20 أغسطس 2020

Launch HN: GitDuck (YC S20) – Zoom for developers with real-time code sharing https://ift.tt/2EcOY8D

Launch HN: GitDuck (YC S20) – Zoom for developers with real-time code sharing Hi everyone! We are Dragos and Thiago from GitDuck ( https://gitduck.com ). We are building GitDuck, a Zoom for developers with direct integration to the IDE so software developers can talk and collaborate in real-time. It all started by accident, Dragos and I were working on something else, a screen recording tool and we started to use it internally to record short videos of our code. At first it was just for quick code reviews and to debug, but soon we realized how helpful it was to have a video explanation of the code. Kind of rubber duck debugging with video. ;) After talking to almost 300 developers and learning that other people were facing similar collaboration issues we decided to focus 100% on building this tool. We are the first users and we use GitDuck internally for quick assistance, pair programming, code reviews or just discussing ideas. It has the features you would expect in a video call tool — like audio, video chat and screen sharing, but the UX and the integrations were built exclusively for developers. You can easily share your code and do pair programming. We are building integrations for all the IDEs. This enables you to collaborate without screen sharing (so it's faster and and consumes less bandwidth), directly from your IDE and independently of the IDE that other people are using. Whenever you join a GitDuck meeting, your IDE extension wakes up and allows you to share your code with the other meeting participants (or join the already shared code from other meeting participant). When your peers join your code, they can see and edit your files in real-time, similar to the Google Docs experience. At any given point you can also go to your peers position so you can see in which file and line they are. Check a 1 min demo ( https://ift.tt/2FDl91n ) GitDuck currently has integrations to VS Code and VSCodium. In the next few days we are going to release the integrations to all JetBrains IDEs. Vim, Sublime and others coming after that. One important aspect to mention is security. We are the first users of the service so we focus a lot on building something that we would trust to use ourselves. All the files shared from your IDE are always shared via peer-to-peer and are end-to-end encrypted. No piece of code never touches our servers, so we never have access to your code. All calls are encrypted and p2p (if 4 or less participants). If 5 or more people join we switch to a cloud infrastructure in order to maintain the quality, but the media are always encrypted and we never have access to your calls. You can read more about it here ( https://ift.tt/3aIT0S1 ) and we are always open for your suggestions to improve. We would love to hear your thoughts and feedback. What are your ideas about tools like this? Thank you! August 20, 2020 at 10:54PM

Launch HN: Hubble (YC S20) – Monitor data quality inside data warehouses https://ift.tt/3gno0Zj

Launch HN: Hubble (YC S20) – Monitor data quality inside data warehouses Hey everyone! We’re Oliver and Hamzah from Hubble ( https://gethubble.io/hn ). Hubble runs tests on your data warehouse so you can identify issues with data quality. You can test for things like missing values, uniqueness of data or how frequently data is added/updated. We worked together for the last 4 years at a startup where we built and managed data products for insurers and banks. A common pattern we saw was teams taking data from their internal tools (CRM, HR system, etc.), application databases, and 3rd party data and storing it in a warehouse for analysis. However, when analysts/data scientists used the data for reports they would spot something suspicious and the engineering team would have to manually go through the data pipelines to find the source of the problem. More often than not it was simple things like a spike in missing values because an ETL job failed or stale data because a 3rd party data source hadn’t updated correctly. We realised that reliability/ trustworthiness of the raw data was essential before you could start abstracting away more interesting tasks like analysis, insight or predictions. We wanted to do this without having to write and maintain lots of individual tests in our code. So we built Hubble, which connects to a data warehouse and creates tests based on the type of data being stored (i.e. freshness of timestamps, the cardinality of strings, max value of numbers, missing values, etc.). We’ve also added the ability to write any custom tests using a built-in SQL editor. All the tests run on a schedule and you’ll get an email or slack alert when they fail. We’re also building webhooks and an Airflow operator so you can run tests immediately after running an ETL job or trigger a process to fix a failing test. Instead of asking users to send their data to us, the tests are run in the data warehouse and we track the test results over time. Today we support BigQuery, Snowflake and Rockset (which lets us work with MongoDB and DynamoDB) and are adding more on request. We’re planning on charging $200 a month for a few seats, and $30-50 for extra users after that. We’re still at an early access stage but want the HN community’s feedback so we’ve opened up access to the app for a few days, you can try it out here https://gethubble.io/hn . We’ve added a demo data warehouse you can start with that has data on COVID-19 cases in Italy and bike-share trips in San Francisco. Thanks and looking forward to hearing your ideas, experiences and feedback! August 20, 2020 at 05:38PM

Show HN: Wishlist – Collect and organize user feedback https://ift.tt/3aOzh3C

Show HN: Wishlist – Collect and organize user feedback As founders, we know how important it is to talk to our customers in order to avoid wasting time building features that no one wants. It can be difficult to know what to work on next, and how many resources to devote to a particular product or feature. That's why I've decided to build https://getwishlist.io, a (currently free in beta) user feedback tool that will help founders like ourselves not only collect user feedback, but also organize it, and build product roadmaps. As such, I'd love to speak with my fellow founders about how you collect feedback from your team and users, how you do your product roadmaps, and some of the challenges you face while doing so. How do you collect and organize feedback from your users? What are your processes? What don’t you like about them? Care to share? August 20, 2020 at 12:36PM